What are the top 10 ways cybercrime and deep fakes and generative AI are exploiting and bullying and potentially killing our children?

What are the top 10 ways cybercrime and deep fakes and generative AI are exploiting and bullying and potentially killing our children?

What are the top 10 ways cybercrime and deep fakes and generative AI are exploiting and bullying and potentially killing our children?

In today’s digital age, children face unprecedented risks from cybercrime, deep fakes, and generative AI. These sophisticated technologies are being misused to exploit, bully, and even endanger young lives. As parents, educators, and guardians, understanding these threats is crucial to safeguarding our children’s online and offline well-being. This blog explores the top 10 ways these digital dangers are impacting our kids and offers insights on how to protect them.

What are the top 10 ways cybercrime, deep fakes, and generative AI are exploiting, bullying and potentially killing our children? Listen here

What are the top 10 ways cybercrime, deep fakes, and generative AI are exploiting, bullying and potentially killing our children?
What are the top 10 ways cybercrime, deep fakes, and generative AI are exploiting, bullying and potentially killing our children?

 

1. Online Predators

Cybercriminals use social media and chat platforms to groom and exploit children, posing as peers to gain their trust.

2. Deep Fake Videos

Manipulated videos can be used to bully or blackmail children, damaging their reputation and causing emotional distress.

3. Cyberbullying

AI-driven bots and fake profiles harass children, leading to severe psychological impacts and even self-harm.

4. Identity Theft

Hackers steal children’s identities to commit fraud, leaving long-lasting financial and legal repercussions.

5. Inappropriate Content

Generative AI can create explicit or harmful content targeting children, exposing them to dangerous material.

6. Online Scams

Children fall victim to phishing scams and fraudulent schemes, often unknowingly providing sensitive information.

7. Privacy Invasion

Smart devices and apps can track and share children’s personal data without consent, compromising their safety.

8. Dark Web Exposure

Children can be lured into dark web activities, where they are exposed to illegal and harmful content.

9. Mental Health Impact

Continuous exposure to online threats can lead to anxiety, depression, and other mental health issues.

10. Physical Danger

In extreme cases, online interactions can lead to real-world abductions or violence, putting children’s lives at risk.

These threats are causing significant concern among law enforcement, child protection agencies, and cybersecurity experts. The rapid advancement of AI technology is making it increasingly difficult to distinguish between real and fake content, putting children at greater risk of exploitation and abuse. Authorities and organizations are calling for increased education, awareness, and legislative measures to combat these growing threats to children’s safety online.

What is CSAM?

CSAM stands for Child Sexual Abuse Material. This term is used to refer to content that visually depicts minors engaged in sexual activity or in a sexually explicit manner. It includes photographs, videos, digital media, and other forms of visual representation of child sexual abuse. The term CSAM is preferred over “child pornography” by law enforcement, child protection agencies, and other professionals working to combat this issue. There are a few key reasons for using this terminology:

  • It emphasizes that the content is a record of child abuse, not a form of pornography.
  • It avoids legitimizing the material by associating it with legal adult pornography.
  • It’s considered more respectful to survivors of child sexual abuse.
  • It accurately describes the criminal nature of the content as evidence of child exploitation and abuse.

CSAM is considered a serious criminal offense in most jurisdictions worldwide and is the subject of international efforts to prevent its production, distribution, and possession. The use of this terminology helps to maintain focus on the gravity of these crimes and the harm done to child victims.

Governments are taking several measures to protect children from AI-generated CSAM:

  1. International Cooperation: The UK and US have pledged to work together to combat AI-generated CSAM. They are committed to developing and funding new capabilities to stop the spread of these images.

  2. Legislative Action: Countries are introducing or updating laws to specifically address AI-generated CSAM. For example, in the US, the PROTECT Act targets the production and distribution of child sexual abuse material, including AI-generated content.

  3. Risk-Based Regulation: Some countries, like the European Union with its proposed AI Act, are adopting risk-based approaches to AI regulation. This ensures that high-risk AI systems, which could potentially be used to create CSAM, are subject to stricter oversight and requirements.

  4. Mandatory Safety Measures for Tech Companies: The UK’s Online Safety Bill requires tech companies to proactively identify and remove CSAM, including AI-generated content. Companies will have a duty to stop the spread of such illegal content.

  5. Empowering Regulatory Bodies: In the UK, Ofcom will have the power to direct companies to use or develop technology to identify and remove CSAM.

  6. Collaborative Efforts: Governments are working with law enforcement agencies, child protection organizations, and tech companies to address the issue. For instance, the UK Home Secretary visited the National Center for Missing and Exploited Children (NCMEC) in the US to discuss joint efforts.

  7. Investing in Detection Technologies: Countries are exploring the development of new solutions to fight the spread of AI-generated CSAM and improve the ability to identify victims and offenders.

  8. Addressing Encryption Concerns: Governments are calling on tech companies to ensure robust safety measures are in place before implementing end-to-end encryption, to maintain the ability to detect child abuse on their platforms.

  9. International Statements and Commitments: Multiple countries and organizations have signed joint statements pledging to work together to tackle child sexual abuse in the age of AI, recognizing the need for international cooperation.

  10. Encouraging Transparency: Governments are urging companies developing frontier AI capabilities to provide transparency on their plans to measure, monitor, and mitigate capabilities that could be exploited by child sexual offenders.

These measures demonstrate a multi-faceted approach to addressing the emerging threat of AI-generated CSAM, combining legislative action, international cooperation, and technological solutions to protect children in the digital age.

AI Innovations in July 2024

AI Innovations in July 2024

AI Innovations in July 2024.

Welcome to our blog series “AI Innovations in July 2024”! As we continue to ride the wave of extraordinary developments from June, the momentum in artificial intelligence shows no signs of slowing down. Last month, we witnessed groundbreaking achievements such as the unveiling of the first quantum AI chip, the successful deployment of autonomous medical drones in remote areas, and significant advancements in natural language understanding that have set new benchmarks for AI-human interaction.

July promises to be just as exhilarating, with researchers, engineers, and visionaries pushing the boundaries of what’s possible even further. In this evolving article, updated daily throughout the month, we’ll dive deep into the latest AI breakthroughs, advancements, and milestones shaping the future.

From revolutionary AI-powered technologies and cutting-edge research to the societal and ethical implications of these innovations, we provide you with a comprehensive and insightful look at the rapidly evolving world of artificial intelligence. Whether you’re an AI enthusiast, a tech-savvy professional, or simply someone curious about the future, this blog will keep you informed, inspired, and engaged.

Join us on this journey of discovery as we explore the frontiers of AI, uncovering the innovations that are transforming industries, enhancing our lives, and shaping our future. Stay tuned for daily updates, and get ready to be amazed by the incredible advancements happening in the world of AI!

LISTEN DAILY AT OUR PODCAST HERE

A  Daily chronicle of AI Innovations July 26th 2024:

🏅AI: The New Gold Medalist in Empowering Athletes at the Olympics

💥 OpenAI challenges Google with AI search engine SearchGPT

🥈 Google DeepMind’s AI takes home silver medal in complex math competition

🎮 Video game actors strike over AI concerns

🚨 Who will control the future of AI?

🏅AI: The New Gold Medalist in Empowering Athletes at the Olympics

AI as a Catalyst for Inclusion

Kevin Piette, paralyzed for 11 years, recently achieved a remarkable milestone by carrying the Olympic flame while walking. This extraordinary feat was made possible by the Atalante X, an AI-powered exoskeleton developed by French company Wandercraft. 🚀

The Olympics have always been a stage for human excellence, a platform where athletes push the boundaries of physical ability. However, the Games are also evolving into a showcase of technological innovation. Artificial intelligence (AI) is rapidly transforming sports, and its impact extends far beyond performance enhancement.

Source: https://etiennenoumen.medium.com/ai-the-new-gold-medalist-in-empowering-athletes-at-the-olympics-c4705500e453

💥 OpenAI challenges Google with AI search engine SearchGPT

  • OpenAI announced a new search product called “SearchGPT,” which is currently in the testing phase and aims to compete directly with Google’s Search Generative Experience.
  • SearchGPT, designed for a limited group of users, offers concise answers and relevant sources, with the intention of making search faster and easier through real-time information.
  • With this move, OpenAI targets Google’s dominant position in the search market, where Google holds approximately 90% market share, highlighting OpenAI’s significant ambition in the search engine space.

Source: https://www.businessinsider.com/openai-searchgpt-search-engine-prototype-declares-war-with-google-2024-5

🥈 Google DeepMind’s AI takes home silver medal in complex math competition

  • Google DeepMind has developed an AI system named AlphaProof that achieved 28 points in the International Mathematical Olympiad, equivalent to a silver medalist’s score for the first time.
  • AlphaProof has managed to solve 83% of all IMO geometry problems over the past 25 years, significantly improving on its predecessor AlphaGeometry, which had a success rate of 53%.
  • AlphaProof generates solutions by searching and testing various mathematical steps, unlike human participants who rely on theorem knowledge and intuition to solve problems more efficiently.

Source: https://www.semafor.com/article/07/25/2024/google-deepminds-ai-reaches-milestone-in-international-mathematical-olympiad

🎮 Video game actors strike over AI concerns 

  • The Screen Actors Guild-American Federation of Television and Radio Artists (SAG-AFTRA) has decided to strike all video game work under the union’s Interactive Media Agreement starting July 26th.
  • The strike affects all union actors, voice actors, and motion capture performers, targeting companies such as Activision Blizzard, EA, Insomniac Games, and WB Games, with disagreements over AI protections cited as the main issue.
  • Despite finding common ground on numerous proposals and the video game producers offering AI consent and fair compensation, SAG-AFTRA and the companies failed to reach a full agreement, leading to the strike.

Source: https://www.theverge.com/2024/7/25/24206357/video-game-performer-strike-sag-aftra

🚨 Who will control the future of AI?

Sam Altman, CEO of OpenAI, just wrote an op-ed outlining a strategy for ensuring a vision for AI prevails in the United States and allied nations over authoritarian alternatives.

  • Altman emphasizes the urgent need for a U.S.-led global coalition to advance AI that spreads its benefits and maintains open access.
  • He proposes four key actions: robust security measures, infrastructure investment, coherent commercial diplomacy, and new models for global AI governance.
  • The strategy aims to maintain the U.S. lead in AI development while countering efforts by authoritarian regimes to dominate the technology.
  • Altman suggests creating an international body for AI oversight, similar to the IAEA or ICANN.

Altman’s surprisingly urgent tone in this op-ed highlights the growing risks of AI development in the US. He believes “there is no third option,” either democratic nations lead AI development or authoritarian regimes will — raising a serious call to action for the race of AI dominance.

Source: https://x.com/sama/status/1816496304257941959

AI video startup Runway reportedly trained on ‘thousands’ of YouTube videos without permission.

Source: https://www.engadget.com/ai-video-startup-runway-reportedly-trained-on-thousands-of-youtube-videos-without-permission-182314160.html


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

Amazon racing to develop AI chips cheaper, faster than Nvidia’s, executives say.

Source: https://www.reuters.com/technology/artificial-intelligence/amazon-racing-develop-ai-chips-cheaper-faster-than-nvidias-executives-say-2024-07-25/

Sam Altman, under fire from Elon Musk, has now offered his own vision of open-source AI.

Source: https://www.businessinsider.com/sam-altman-under-fire-elon-musk-vision-open-source-ai-2024-7

Gemini is now 20% faster than OpenAI’s most advanced model.

Source: https://www.newsbytesapp.com/news/science/google-s-gemini-gets-speed-boost-with-new-1-5-flash-model/story

JP Morgan built its own AI chatbot that acts like a ‘research analyst’.

Source: https://decrypt.co/241834/jp-morgan-ai-chatbot-llm-suite

Google upgraded Gemini with 1.5 Flash, offering faster responses, a 4x larger context window, and expanded access in over 40 languages and 230 countries.

Source: https://blog.google/products/gemini/google-gemini-new-features-july-2024/

SAG-AFTRA announced a strike for video game performers starting July 26, citing concerns over AI protections in negotiations with major gaming studios, despite progress on wages and job safety.

Source: https://apnews.com/article/sagaftra-video-game-performers-ai-strike-4f4c7d846040c24553dbc2604e5b6034

Sam Altman revealed in a tweet reply that the GPT-4o-Voice Alpha rollout will begin next week for Plus subscribers, expanding OpenAI’s voice generation capabilities.

Source: https://x.com/sama/status/1816560608554418401

Udio released version 1.5 of its AI music model, featuring improved audio quality, key control, and new features like stem downloads and audio-to-audio remixing.

Source: https://www.udio.com/blog/introducing-v1-5

Runway’s AI video generator reportedly trained on thousands of YouTube videos without permission, according to a leaked document obtained by 404 Media.

Source: https://www.404media.co/runway-ai-image-generator-training-data-youtube

Anthropic’s web crawler allegedly violated website terms of use, with iFixit reporting nearly a million hits in 24 hours, raising concerns about AI companies’ data collection practices.

Source: https://www.theverge.com/2024/7/25/24205943/anthropic-ai-web-crawler-claudebot-ifixit-scraping-training-data

A  Daily chronicle of AI Innovations July 25th 2024:

💸 OpenAI could lose $5B this year and run out of cash in 12 months

🎥 Kling AI’s video generation goes global

🗺️ Apple Maps launches on the web to take on Google

🚨 Mistral’s Large 2 is its answer to Meta and OpenAI’s latest models

🙃 CrowdStrike offers $10 Uber Eats gift cards as an apology for the outage

👀 Reddit blocking all search engines except Google, as it implements AI paywall

🇫🇷 Mistral’s Large 2 takes on AI giants

💸 OpenAI could lose $5B this year and run out of cash in 12 months

  • OpenAI could lose up to $5 billion in 2024, risking running out of cash within 12 months, according to an analysis by The Information.
  • The AI company is set to spend $7 billion on artificial intelligence training and $1.5 billion on staffing this year, far exceeding the expenses of rivals.
  • OpenAI may need to raise more funds within the next year to sustain its operations, despite having already raised over $11 billion through multiple funding rounds.

Source: https://cointelegraph.com/news/openai-could-lose-5b-this-year-and-run-out-of-cash-in-12-months-report

🚨 Mistral’s Large 2 is its answer to Meta and OpenAI’s latest models

  • French AI company Mistral AI launched its Mistral Large 2 language model just one day after Meta’s release of Llama 3, highlighting the intensifying competition in the large language model (LLM) market.
  • Mistral Large 2 aims to set new standards in performance and efficiency, boasting significant improvements in logic, code generation, and multi-language support, with a particular focus on minimizing hallucinations and improving reasoning capabilities.
  • The model, available on multiple platforms including Azure AI Studio and Amazon Bedrock, outperforms its predecessor with 123 billion parameters and supports extensive applications, signaling a red ocean of competition in the AI landscape.

Source: https://the-decoder.com/mistral-large-2-just-one-day-after-llama-3-signals-the-llm-market-is-getting-redder-by-the-day/

👀 Reddit blocking all search engines except Google, as it implements AI paywall

  • Reddit has begun blocking search engines from accessing recent posts and comments, except for Google, which has a $60 million agreement to train its AI models using Reddit’s content.
  • This move is part of Reddit’s strategy to monetize its data and protect it from being freely used by popular search engines like Bing and DuckDuckGo.
  • To enforce this policy, Reddit updated its robots.txt file, signaling to web crawlers without agreements that they should not access Reddit’s data.

Source: https://www.theverge.com/2024/7/24/24205244/reddit-blocking-search-engine-crawlers-ai-bot-google

🎥 Kling AI’s video generation goes global

Kling AI, developed by Chinese tech giant Kuaishou Technology, has released its impressive AI video model globally, offering high-quality AI generations that rival OpenAI’s (unreleased) Sora.

  • Kling can generate videos up to two minutes long, surpassing OpenAI’s Sora’s one-minute limit, however, the global version is limited to five-second generations.
  • The global version offers 66 free credits daily, with each generation costing 10 credits.
  • According to Kuaishou, Kling utilizes advanced 3D reconstruction technology for more natural movements.
  • The platform accepts prompts of up to 2,000 characters, allowing for detailed video descriptions.

When KLING launched a little over a month ago, it was only accessible if you had a Chinese phone number. While global users are still limited to 5-second generations, anyone can now generate their own high-quality videos — putting even more pressure on OpenAI to release its beloved Sora.

Source: https://klingai.com/

Stability AI introduces Stable Video 4D, its new AI model for 3D video generation.

Source: https://siliconangle.com/2024/07/24/stability-ai-introduces-stable-video-4d-new-ai-model-3d-video-generation/

Microsoft is adding AI-powered summaries to Bing search results.

Source: https://www.engadget.com/microsoft-is-adding-ai-powered-summaries-to-bing-search-results-203053790.html

👀 OpenAI unveils SearchGPT

OpenAI, whose ChatGPT assistant kicked off an artificial intelligence arms race, is now pursuing a slice of the search industry. The company has unveiled a prototype of SearchGPT, an AI-powered search engine that is widely viewed as a play for rival Google’s $175 billion-per-year search business. But while Google’s use of AI in search results has been met with concern and resistance from publishers, SearchGPT touts its heavy use of citations and was developed alongside publishing partners, including Axel-Springer and the Financial Times. After seeing results to their queries, users will be able to ask follow-up questions in interactions that resemble those with ChatGPT.

  • A 10,000 person wait list was opened Thursday for a those wanting to test a prototype of the SearchGPT service.
  • Though currently distinct, SearchGPT will eventually be integrated into ChatGPT.

Source: chatgpt.com

A  Daily chronicle of AI Innovations July 24th 2024:

📈 Google search is thriving despite AI shift

🚗 Google is pouring billions into self-driving taxis as Tesla prepares to reveal its rival

🚨 Senators demand answers on OpenAI’s practices

🦙 Meta’s Llama 3.1 takes on GPT-4o

🔥 Adobe’s new AI features for Photoshop

📈 Google search is thriving despite AI shift 

  • Despite concerns from online publishers, Google’s introduction of AI features generating conversational responses to search queries has attracted advertisers and propelled Alphabet’s success.
  • Alphabet’s revenue for the April-June quarter rose by 14% from last year to $84.74 billion, surpassing analyst expectations and boosting stock prices by 2% in extended trading.
  • Google’s cloud-computing division, its fastest-growing segment, generated $10.3 billion in revenue in the past quarter, marking its first time surpassing the $10 billion threshold in a single quarter.

Source: https://www.fastcompany.com/91161798/google-search-is-still-thriving-despite-a-shift-to-ai-earnings

🚗 Google is pouring billions into self-driving taxis as Tesla prepares to reveal its rival

  • Alphabet is investing $5 billion in Waymo’s self-driving taxi service, highlighting its commitment to autonomous vehicles.
  • Waymo has achieved over 50,000 paid autonomous rides weekly in cities like San Francisco and Phoenix, showcasing its progress and customer acceptance.
  • Tesla is also preparing to enter the self-driving taxi market, with an important event unveiling its rival service rescheduled from August to October.

Source: https://www.businessinsider.com/alphabet-is-pouring-billions-into-waymos-self-driving-vehicles-2024-7

🚨 Senators demand answers on OpenAI’s practices

Five U.S. Senators have just sent a letter to OpenAI CEO Sam Altman, demanding details about the company’s efforts to ensure AI safety following reports of rushed safety testing for GPT-4 Omni.

  • Senators question OpenAI’s safety protocols, citing reports that the company rushed safety testing of GPT-4 Omni to meet a May release date.
  • The letter requests OpenAI to make its next foundation model available to U.S. Government agencies for deployment testing, review, analysis, and assessment.
  • Lawmakers ask if OpenAI will commit 20% of computing resources to AI safety research, a promise made in July 2023 when announcing the now disbanded “Superalignment team”.

With allegations of rushed safety testing, potential retaliation against whistleblowers, and the disbanding of the “Superalignment team,” OpenAI is under intense scrutiny. This letter also marks a critical moment for the entire AI industry — with the potential to lead to stricter government oversight and new industry standards.

Source: https://cointelegraph.com/news/us-lawmakers-letter-open-ai-requesting-government-access

🦙 Meta’s Llama 3.1 takes on GPT-4o

Ace the Microsoft Azure Fundamentals AZ-900 Certification Exam: Pass the Azure Fundamentals Exam with Ease

In case you missed our exclusive deep dive with Mark Zuckerberg yesterday, Meta released Llama 3.1, including it’s long awaited 405B paramater model — the first open sourced frontier model that beats top closed models like GPT-4o across several benchmarks.

  • The 405B parameter version of Llama 3.1 matches or exceeds top closed models on several benchmarks.
  • Meta is offering open and free weights and code, with a license enabling fine-tuning, distillation into other models, and deployment anywhere.
  • Llama 3.1 features a 128k context length, multi-lingual abilities, strong code generation performance, and complex reasoning capabilities.
  • For exclusive insights on Llama 3.1, open source, AI agents, and more, read our full deep dive with Mark Zuckerberg here, or watch the full interview here.

Meta’s release of Llama 3.1 405b is a significant moment in AI history because it’s the first time an open-source AI model matches or outperforms top closed AI models like OpenAI’s GPT-4o. By offering a private, customizable alternative to closed AI systems, Meta is enabling anyone to create their own tailored AI.

Source: https://www.therundown.ai/p/meta-releases-llama-405b

🔥 Adobe’s new AI features for Photoshop

Adobe just unveiled major AI-powered updates to Illustrator and Photoshop, leveraging its Firefly AI model to accelerate creative workflows and introduce new generative design capabilities.

  • Illustrator introduces Generative Shape Fill using Firefly Vector AI to add detailed vectors to shapes and create scalable patterns via text prompts.
  • Text to Pattern in Illustrator creates scalable, customized vector patterns for designs like wallpapers.
  • Photoshop’s new AI-powered Selection Brush Tool and Generate Image function are now generally available.
  • Photoshop also gets an enhanced version of its popular Generative Fill for improved sharpness in large images.

These updates could dramatically increase designers’ productivity by automating tedious, time-consuming tasks. We’ve always preached that the best AI products are those embedded into everyday workflows — and Adobe is doing just that by putting powerful tech directly into designers’ everyday tools.

Source: https://news.adobe.com/news/news-details/2024/Adobe-Unveils-Powerful-New-Innovations-in-Illustrator-and-Photoshop-Unlocking-New-Design-Possibilities-for-Creative-Pros/default.aspx

Mark Zuckerberg explains why open source AI is good for developers.

Source: https://www.neowin.net/news/mark-zuckerberg-explains-why-open-source-ai-is-good-for-developers/

Google has big new ideas about the Play Store.

The company is rolling out several new features including Collections, AI-powered app comparisons, and more

Source: https://www.theverge.com/2024/7/24/24205052/google-play-collections-ai-features-rewards-pixel

OpenAI offers free GPT-4o Mini fine-tuning to counter Meta’s Llama 3.1 release.

Source: https://venturebeat.com/ai/ai-arms-race-escalates-openai-offers-free-gpt-4o-mini-fine-tuning-to-counter-metas-llama-3-1-release/

A  Daily chronicle of AI Innovations July 23rd 2024:

🔮 Meta releases its most powerful AI model yet

💸 Alexa is losing Amazon billions of dollars

🚀 The “world’s most powerful” supercomputer

🌦️ Google’s AI-powered weather model

🧬 MIT’s AI identifies breast cancer risk

🔋 Musk unveils the world’s most powerful AI training cluster
🤖 Robotics won’t have a ChatGPT-like explosion: New Research
🌦️ NeuralGCM predicts weather faster than SOTA climate models

 
🤖 Robotics won’t have a ChatGPT-like explosion: New Research

Coatue Management has released a report on AI humanoids and robotics’s current and future state. It says robotics will unlikely have a ChatGPT-like moment where a single technology radically transforms our work. While robots have been used for physical labor for over 50 years, they have grown linearly and faced challenges operating across different environments.

The path to broad adoption of general-purpose robots will be more gradual as capabilities improve and costs come down. Robotics faces challenges like data scarcity and hardware limitations that digital AI technologies like ChatGPT do not face. But investors are still pouring billions, hoping software innovations could help drive value on top of physical robotics hardware.

Why does it matter?

We’re on the cusp of a gradual yet profound transformation. While robotics may not suddenly become ubiquitous, the ongoing progress in artificial intelligence and robotics will dramatically alter the landscape of numerous fields, including manufacturing and healthcare.

Source: https://www.coatue.com/blog/perspective/robotics-wont-have-a-chatgpt-moment

🌦️ NeuralGCM predicts weather faster than SOTA climate models

Google researchers have developed a new climate modeling tool called NeuralGCM. This tool uses a combination of traditional physics-based modeling and machine learning. This hybrid approach allows NeuralGCM to generate accurate weather and climate predictions faster and more efficiently than conventional climate models.

If you are looking for an all-in-one solution to help you prepare for the AWS Cloud Practitioner Certification Exam, look no further than this AWS Cloud Practitioner CCP CLF-C02 book

NeuralGCM’s weather forecasts match the accuracy of current state-of-the-art (SOTA) models for up to 5 days, and its ensemble forecasts for 5-15 day predictions outperform the previous best models. Additionally, NeuralGCM’s long-term climate modeling is one-third as error-prone as existing atmosphere-only models when predicting temperatures over 40 years.

Why does it matter?

NeuralGCM presents a new approach to building climate models that could be faster, less computationally costly, and more accurate than existing models. This breakthrough could lead to accessible and actionable climate modeling tools.

Source: https://research.google/blog/fast-accurate-climate-modeling-with-neuralgcm

🚀 The “world’s most powerful” supercomputer

Elon Musk and xAI just announced the Memphis Supercluster — “the most powerful AI training cluster in the world“, also revealing that Grok 3.0 is planned to be released in December and should be the most powerful AI in the world.

  • Musk tweeted that xAI just launched the “Memphis Supercluster,” using 100,000 Nvidia H100 GPUs, making it “the most powerful AI training cluster in the world.”
  • The xAI founder also revealed that Grok 2.0 is done training and will be released soon.
  • The supercluster aims to create the “world’s most powerful AI by every metric”, Grok 3.0, by December 2024.
  • In a separate tweet yesterday, Musk also revealed that Tesla plans to have humanoid robots in “low production” for internal use next year.

 Love him or hate him, the speed at which Elon and the team at xAI operate has been wild to witness. If estimates are accurate, xAI might be on track to create the most powerful AI systems in the world by year’s end — solidifying its position as one of the top competitors in the space and not just another AI startup.

Source: https://x.com/elonmusk/status/1815325410667749760

🌦️ Google’s AI-powered weather model

Google researchers have developed a new AI-powered weather and climate model called ‘NeuralGCM’ by combining methods of machine learning and neural networks with traditional physics-based modeling.

  • NeuralGCM has proven more accurate than purely machine learning-based models for 1-10 day forecasts and top extended-range models.
  • NeuralGCM is up to 100,000 times more efficient than other models for simulating the atmosphere.
  • The model is open-source and can run relatively quickly on a laptop, unlike traditional models that require supercomputers.

At up to 100,000 times more efficient than traditional models — NeuralGCM could dramatically enhance our ability to simulate complex climate scenarios quickly and accurately. While still a ton of adoption challenges ahead, it’s a big leap forward for more informed climate action and resilience planning.

Source: https://www.nature.com/articles/s41586-024-07744-y

🧬 MIT’s AI identifies breast cancer risk

The Rundown: Researchers from MIT and ETH Zurich have developed an AI model that can identify different stages of ductal carcinoma in situ (DCIS), a type of preinvasive breast tumor, using simple tissue images.

  • The model analyzes chromatin images from 560 tissue samples (122 patients), identifying 8 distinct cell states across DCIS stages.
  • It considers both cellular composition and spatial arrangement, revealing that tissue organization is crucial in predicting disease progression.
  • Surprisingly, cell states associated with invasive cancer were detected even in seemingly normal tissue.

This AI model could democratize advanced breast cancer diagnostics, offering a cheaper, faster way to assess DCIS risk. While clinical validation is still needed, AI is likely going to work hand-in-hand with pathologists in the near future to catch cancer earlier and more accurately.

Source: https://www.nature.com/articles/s41467-024-50285-1

🔮 Meta releases its most powerful AI model yet

  • Meta has released Llama 3.1 405B, its largest open-source AI model to date, featuring 405 billion parameters which enhance its problem-solving abilities.
  • Trained with 16,000 Nvidia H100 GPUs, Llama 3.1 405B is competitive with leading AI models like OpenAI’s GPT-4 and Anthropic’s Claude 3.5 Sonnet, though it has specific strengths and weaknesses.
  • Meta’s new AI model is available for download or cloud usage and powers chatbots on platforms like WhatsApp and Meta.ai, showcasing capabilities in coding, mathematical queries, and multilingual document summarization.

Source: https://techcrunch.com/2024/07/23/meta-releases-its-biggest-open-ai-model-yet/

💸 Alexa is losing Amazon billions of dollars

  • Amazon plans to launch a paid version of Alexa to address the over $25 billion losses incurred by its devices business from 2017 to 2021, as reported by The Wall Street Journal.
  • The enhanced Alexa, which may cost up to $10 per month, is expected to be released soon, though employees have concerns about whether the technology is ready.
  • The new Alexa, featuring generative AI for improved conversational abilities, faces technical delays and competition from free AI assistants, raising doubts about customers’ willingness to pay for it.

Source: https://www.theverge.com/2024/7/23/24204260/amazon-25-billion-losses-echo-devices-alexa-subscription

What Else Is Happening in AI on July 23rd 2024❗

💊 VeriSIM Life’s AI platform can accelerate drug discovery

VeriSIM Life has developed an AI platform, BIOiSIM, to help speed up drug discovery and reduce animal testing. The platform contains data on millions of compounds and uses AI models to predict how potential new drugs will work in different species, including humans.

Source: https://venturebeat.com/ai/can-ai-increase-the-pace-and-quality-of-pharmaceutical-research-verisim-life-says-yes

📷 Anthropic is working on a new screenshot tool for Claude

This tool will allow users to capture and share screenshots from their desktop or browser directly within the Claude chat interface. It will streamline the sharing of visual information and code snippets when asking Claude for assistance on tasks like coding or troubleshooting.

Source: https://www.testingcatalog.com/anthropic-working-on-new-screenshot-tool-for-claude-ai/

🔂 Luma’s “Loops” feature in Dream Machine transforms digital marketing

The “Loops” feature allows users to create continuous video loops from text descriptions or images. It does so without visible cuts or transitions, opening up new possibilities for engaging content creation and advertising.

Source: https://venturebeat.com/ai/how-luma-ais-new-loops-feature-in-dream-machine-could-transform-digital-marketing

🤖 Tesla will use humanoid robots internally by next year

Elon Musk has announced that Tesla will use humanoid robots at its factories by next year. These robots, called Optimus, were expected to be ready by the end of 2024. Tesla aims to mass produce robots for $20,000 each and sell them to other companies starting in 2026.

Source: https://www.reuters.com/business/autos-transportation/tesla-have-humanoid-robots-internal-use-next-year-musk-says-2024-07-22

🎤 Perplexity launches Voice Mode for its AI assistant on iOS

Perplexity has introduced a new feature for its iOS app called Voice Mode. It allows subscribers with Pro accounts to interact verbally with the AI-powered search engine. Users can now engage in voice-based conversations and pose questions using various voice options.

Source: https://x.com/perplexity_ai/status/1814348871746585085

A  Daily chronicle of AI Innovations July 22nd 2024:

🤖 Apple released two open-source AI language models
🤝 OpenAI is in talks with Broadcom to develop an AI chip
🖥️ Nvidia is developing an AI chip series for China

🤖 The state of AI humanoids and robotics

🍎 Apple’s new 7B open-source AI model

🤖 Tesla to have humanoid robots for internal use next year

🇨🇳 Nvidia preparing new flagship AI chip for Chinese market

⚡️ Musk’s xAI turns on ‘world’s most powerful’ AI training cluster

📈 Study reveals rapid increase in web domains blocking AI models

⚙️ How to test and customize GPT-4o mini

🤖 Apple released two open-source AI language models

Apple has released two new open AI models called DCLM (DataComp for Language Models) on Hugging Face: one with 7 billion parameters and another with 1.4 billion parameters. The 7B model outperforms Mistral-7B and is comparable to other leading open models, such as Llama 3 and Gemma. They’ve released – model weights, training code, and even the pretraining dataset. The models were trained using a standardized framework to determine the best data curation strategy.

Source: https://venturebeat.com/ai/apple-shows-off-open-ai-prowess-new-models-outperform-mistral-and-hugging-face-offerings

The 7B model was trained on 2.5 trillion tokens and has a 2K context window, achieving 63.7% 5-shot accuracy on MMLU. The 1.4B model, trained on 2.6 trillion tokens, outperforms other models in its category on MMLU with a score of 41.9%. These models are not intended for Apple devices.

Why does it matter?

By open-sourcing high-performing models and sharing data curation strategies, Apple is helping to solve some of AI’s toughest challenges for developers and researchers. This could lead to more efficient AI applications across various industries, from healthcare to education.

Source: https://venturebeat.com/ai/apple-shows-off-open-ai-prowess-new-models-outperform-mistral-and-hugging-face-offerings

🤝 OpenAI is in talks with Broadcom to develop an AI chip

The company is in talks with Broadcom and other chip designers to build custom silicon, aiming to reduce dependence on Nvidia’s GPUs and boost its AI infrastructure capacity. OpenAI is hiring ex-Google employees with AI chip experience and has decided to develop an AI server chip.

The company is researching various chip packaging and memory components to optimize performance. However, the new chip is not expected to be produced until 2026 at the earliest.

Why does it matter?

Sam Altman’s vision for AI infrastructure is evolving from a separate venture into an in-house project at OpenAI. By bringing chip design in-house, OpenAI could potentially accelerate its AI research, reduce dependencies on external suppliers, and gain a competitive edge in the race of advanced AI.

Source: https://www.theinformation.com/articles/openai-has-talked-to-broadcom-about-developing-new-ai-chip

🖥️ Nvidia is developing an AI chip series for Chi

Nvidia is developing a special version of its Blackwell AI chip for the Chinese market. Tentatively named “B20,” this chip aims to bridge the gap between U.S. export controls and China’s AI tech. Despite facing a revenue dip from 26% to 17% in China due to sanctions, Nvidia is not backing down. They’re partnering with local distributor Inspur to launch this new chip.

As Nvidia tries to reclaim its Chinese market share, competitors like Huawei are gaining ground. Meanwhile, the U.S. government is making even tighter controls on AI exports.

Why does it matter?

If Nvidia pulls off, it could maintain its dominance in the Chinese market while complying with U.S. regulations. But if regulators clamp down further, we could see a more fragmented global AI ecosystem, potentially slowing innovation. It’s a high-stakes game of technological cat-and-mouse, with Nvidia trying to stay ahead of regulators and rivals.

Source: https://www.reuters.com/technology/nvidia-preparing-version-new-flaghip-ai-chip-chinese-market-sources-say-2024-07-22

🤖 Tesla to have humanoid robots for internal use next year 

  • Elon Musk announced that Tesla’s Optimus robots will begin “low production” for internal tasks in 2025, with mass production for other firms starting in 2026.
  • Musk initially stated the Optimus robot would be ready to perform tasks in Tesla’s EV factories by the end of this year.
  • Musk’s plans for Optimus and AI products come as Tesla faces reduced demand for electric vehicles and anticipates low profit margins in upcoming quarterly results.

Source: https://www.newsbytesapp.com/news/science/tesla-s-optimus-humanoid-robots-set-for-internal-use-by-2025/story

⚡Musk’s xAI turns on ‘world’s most powerful’ AI training cluster

  • Elon Musk’s xAI has started training its AI models using over 100,000 Nvidia H100 GPUs at a new supercomputing facility in Memphis, Tennessee, described as the most powerful AI training cluster globally.
  • This facility, known as the “Gigafactory of Compute,” is built in a former manufacturing site, and xAI secured $6 billion in funding, creating jobs for roles like fiber foreman, network engineer, and project manager.
  • The Memphis supercomputing site’s large energy and water demands have raised concerns among local environmental groups and residents, who fear its significant impact on water supplies and electrical consumption.

Source: https://www.pcmag.com/news/elon-musk-xai-powers-up-100k-nvidia-gpus-to-train-grok

📈 Study reveals rapid increase in web domains blocking AI models 

  • A new study finds that more websites are blocking AI models from accessing their training data, potentially leading to less accurate and more biased AI systems.
  • The Data Provenance Initiative conducted the study, analyzing 14,000 web domains and discovering an increase in blocked tokens from 1% to up to 7% from April 2023 to April 2024.
  • News websites, social media platforms, and forums are the primary sources of these restrictions, with blocked tokens on news sites rising dramatically from 3% to 45% within a year.

Source: https://the-decoder.com/study-reveals-rapid-increase-in-web-domains-blocking-ai-models-from-training-data/

What Else Is Happening in AI on July 22nd 2024❗

📰 The Reuters Institute released a study on public attitudes about AI in the news

It indicates that news consumers aren’t gloomy about AI in journalism. While initial reactions tend to be skeptical, attitudes become more nuanced as people learn about different AI applications. The comfort level varies based on where AI is used in the news process, with human oversight remaining a top priority.

Source: https://reutersinstitute.politics.ox.ac.uk/news/ok-computer-understanding-public-attitudes-towards-uses-generative-ai-news

🚨California pushes bill requiring tech giants to test AI for “catastrophic” risks

While Republicans pledge a hands-off approach nationally, California’s move has sparked fierce debate. Tech leaders oppose the bill, citing potential harm to innovation and startups, while supporters argue it’s crucial for public safety.

Source: https://www.washingtonpost.com/technology/2024/07/19/biden-trump-ai-regulations-tech-industry

🎨 Figma pulled its “Make Designs” AI tool after it generated designs similar to Apple’s weather app

The design platform admits it rushed new components without proper vetting, leading to uncanny similarities. While Figma didn’t train the AI on copyrighted designs, it’s back to the drawing board to polish its QA process.

Source: https://www.theverge.com/2024/7/18/24201308/figma-make-designs-vet-apple

🛡️ OpenAI’s GPT-4o Mini has a safety feature called “instruction hierarchy”

This new feature prevents users from tricking the AI with sneaky commands like “ignore all previous instructions.” By prioritizing the developer’s original prompts, OpenAI aims to make its AI more trustworthy and safer for future applications, like running your digital life.

Source: https://www.theverge.com/2024/7/19/24201414/openai-chatgpt-gpt-4o-prompt-injection-instruction-hierarchy

🏅 Google is the “official AI sponsor for Team USA” for the 2024 Paris Games

NBCUniversal’s broadcast will feature Google’s tech, from 3D venue tours to AI-assisted commentary. Moreover, Five Olympic and Paralympic athletes will appear in promos using Google’s AI tools.

Source: https://www.theverge.com/2024/7/18/24201440/google-paris-2024-olympic-games-ai-gemini-ads-sponsor

A  Daily chronicle of AI Innovations July 20th 2024:

🍓 OpenAI is working on an AI codenamed “Strawberry”
🧠 Meta researchers developed “System 2 distillation” for LLMs
🛒 Amazon’s Rufus AI is now available in the US
💻 AMD amps up AI PCs with next-gen laptop chips
🎵 YT Music tests AI-generated radio, rolls out sound search
🤖 3 mysterious AI models appear in the LMSYS arena
📅 Meta’s Llama 3 400B drops next week
🚀 Mistral AI adds two new models to its growing family of LLMs
⚡ FlashAttention-3 enhances computation power of NVIDIA GPUs
🏆 DeepL’s new LLM crushes GPT-4, Google, and Microsoft 
🆕 Salesforce debuts Einstein service agent
👨‍🏫 Ex-OpenAI researcher launches AI education company
🔍 OpenAI introduces GPT-4o mini, its most affordable model
🤝 Mistral AI and NVIDIA collaborate to release a new model
🌐 TTT models might be the next frontier in generative AI

🙃 CrowdStrike fixes start at “reboot up to 15 times” and get more complex from there

🍎 Apple releases the “best-performing” open-source models out there

👓 Google in talks with Ray-Ban for AI smart glasses

🚫 Loophole that helps you identify any bot blocked by OpenAI

🍎 Apple releases the “best-performing” open-source models out there

  • Apple’s research team has released open DCLM models on Hugging Face, featuring 7 billion and 1.4 billion parameters, outperforming Mistral and approaching the performance of Llama 3 and other leading models.
  • The larger 7B model achieved a 6.6 percentage point improvement on the MMLU benchmark compared to previous state-of-the-art models while using 40% less compute for training, matching closely with top models like Google’s Gemma and Microsoft’s Phi-3.
  • Currently, the larger model is available under Apple’s Sample Code License, while the smaller one has been released under Apache 2.0, allowing for commercial use, distribution and modification.

Source: https://venturebeat.com/ai/apple-shows-off-open-ai-prowess-new-models-outperform-mistral-and-hugging-face-offerings/

👓 Google in talks with Ray-Ban for AI smart glasses

  • Google is in discussions with EssilorLuxottica, the parent company of Ray-Ban, to develop AI-powered Gemini smart glasses and integrate their Gemini AI assistant.
  • EssilorLuxottica is also collaborating with Meta on the Ray-Ban Meta Smart Glasses, and Meta may acquire a minority stake in EssilorLuxottica, which could affect Google’s plans.
  • Google’s Gemini smart glasses are expected to feature a microphone, speaker, and camera without displays, aligning with the prototypes shown at I/O 2024 for Project Astra.

Source: https://www.newsbytesapp.com/news/science/google-seeks-partnership-with-essilorluxottica-for-smart-glasses-development/story

🚫 Loophole that helps you identify any bot blocked by OpenAI

  • OpenAI developed a technique called “instruction hierarchy” to prevent misuse of AI by ensuring the model follows the developer’s original instructions rather than user-injected prompts.
  • The first model to include this new safety feature is GPT-4o Mini, which aims to block the “ignore all previous instructions” loophole that could be used to exploit the AI.
  • This update is part of OpenAI’s efforts to enhance safety and regain trust, as the company faces ongoing concerns and criticisms about its safety practices and transparency.

Source: https://www.theverge.com/2024/7/19/24201414/openai-chatgpt-gpt-4o-prompt-injection-instruction-hierarchy

A  Daily chronicle of AI Innovations July 19th 2024:

🤖 OpenAI discusses new AI chip with Broadcom

🔮 Mistral AI and Nvidia launch NeMo 12B

🤝 Tech giants form Coalition for Secure AI

🚀OpenAI debuts new GPT-4o mini model

🚀 Mistral AI and NVIDIA collaborate to release a new model
⚡ TTT models might be the next frontier in generative AI

🔓OpenAI gives customers more control over ChatGPT Enterprise

🤝AI industry leaders have teamed up to promote AI security

📈DeepSeek open-sources its LLM ranking #1 on the LMSYS leaderboard

🏆Groq’s open-source Llama AI model tops GPT-4o and Claude

🗣️Apple, Salesforce break silence on claims they used YouTube videos to train AI

🚀OpenAI debuts new GPT-4o mini model

OpenAI just announced the launch of GPT-4o mini, a cost-efficient and compact version of its flagship GPT-4o model — aimed at expanding AI accessibility for developers and businesses.

  • GPT-4o mini is priced at 15 cents per million input tokens and 60 cents per million output tokens, over 60% cheaper than GPT-3.5 Turbo.
  • The model scores 82% on the MMLU benchmark, outperforming Google’s Gemini Flash (77.9%) and Anthropic’s Claude Haiku (73.8%).
  • GPT-4o mini is replacing GPT-3.5 Turbo in ChatGPT for Free, Plus, and Team users starting today.
  • The model supports a 128K token context window and handles text and vision inputs, with audio and video capabilities planned for future updates.

While it’s not GPT-5, the price and capabilities of this mini-release significantly lower the barrier to entry for AI integrations — and marks a massive leap over GPT 3.5 Turbo. With models getting cheaper, faster, and more intelligent with each release, the perfect storm for AI acceleration is forming.

Source: https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence

💪Mistral and Nvidia drop small AI powerhouse

Mistral AI and Nvidia just unveiled Mistral NeMo, a new open-source, 12B parameter small language model that surpasses competitors like Gemma 2 9B and Llama 3 8B on key benchmarks alongside a massive context window increase.

  • NeMo features a 128k token context window, and offers SOTA performance in reasoning, world knowledge, and coding accuracy for its size category.
  • The model also excels in multi-turn conversations, math, and common sense reasoning, making it versatile for various enterprise applications.
  • Mistral also introduced ‘Tekken’, a tokenizer that represents text more efficiently across 100+ languages, allowing for 30% more content within the context window.
  • NeMo is designed to run on a single NVIDIA L40S, GeForce RTX 4090, or RTX 4500 GPU, bringing powerful AI capabilities to standard business hardware.

Small language models are having a moment — and we’re quickly entering a new shift toward AI releases that don’t sacrifice power for size and speed. Mistral also continues its impressive week of releases, continuing to flex the open-source muscle and compete with the industry’s giants.

Source: https://mistral.ai/news/mistral-nemo

⚒️ Groq’s new AI models surge up leaderboard

AI startup Groq just released two new open-source AI models specializing in tool use, surpassing heavyweights like GPT-4 Turbo, Claude 3.5 Sonnet, and Gemini 1.5 Pro on key function calling benchmarks.

  • Groq’s two models, Llama 3 Groq Tool Use 8B and 70B, are both fine-tuned versions of Meta’s Llama 3.
  • The 70B achieved 90.76% accuracy on the BFCL Leaderboard, securing the top position for all proprietary and open-source models.
  • The smaller 8B model was not far behind, coming in at No. 3 on the leaderboard with 89.06% accuracy.
  • The models were trained exclusively on synthetic data, and are available through the Groq API and on Hugging Face.

Groq made waves earlier this year with its blazing-fast AI speeds — and now its pairing those capabilities with top-end specialized models. Near real-time speeds and highly-advanced tool use opens the door for a near endless supply of new innovations and user applications.

Source: https://wow.groq.com/introducing-llama-3-groq-tool-use-models/

🤖 OpenAI introduces GPT-4o mini, its most affordable model

OpenAI has introduced GPT-4o mini, its most intelligent, cost-efficient small model. It supports text and vision in the API, with support for text, image, video and audio inputs and outputs coming in the future. The model has a context window of 128K tokens, supports up to 16K output tokens per request, and has knowledge up to October 2023.

GPT-4o mini scores 82% on MMLU and currently outperforms GPT-4 on chat preferences in the LMSYS leaderboard. It is more affordable than previous frontier models and more than 60% cheaper than GPT-3.5 Turbo.

Why does it matter?

It has been a huge week for small language models (SLMs), with GPT-4o mini, Hugging Face’s SmolLM, and NeMO, Mathstral, and Codestral Mamba from Mistral. GPT-4o mini should significantly expand the range of applications built with AI by making intelligence much more affordable.

Source: https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence

🚀 Mistral AI and NVIDIA collaborate to release a new model

Mistral releases Mistral NeMo, its new best small model with a large context window of up to 128k tokens. It was built in collaboration with NVIDIA and released under the Apache 2.0 license.

Its reasoning, world knowledge, and coding accuracy are state-of-the-art in its size category. Relying on standard architecture, Mistral NeMo is easy to use and a drop-in replacement for any system using Mistral 7B. It is also on function calling and is particularly strong in English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi.

Why does it matter?

The model is designed for global, multilingual applications with excellence in many languages. This could be a new step toward bringing frontier AI models to everyone’s hands in all languages that form human culture.

Source: https://mistral.ai/news/mistral-nemo

⚡ TTT models might be the next frontier in generative AI

Transformers have long been the dominant architecture for AI, powering OpenAI’s Sora, GPT-4o, Claude, and Gemini. But they aren’t especially efficient at processing and analyzing vast amounts of data, at least on off-the-shelf hardware.

Researchers at Stanford, UC San Diego, UC Berkeley, and Meta proposed a promising new architecture this month. The team claims that Test-Time Training (TTT) models can not only process far more data than transformers but that they can do so without consuming nearly as much compute power. Here is the full research paper.

Why does it matter?

On average, a ChatGPT query needs nearly 10x as much electricity to process as a Google search. It may be too early to claim if TTT models will eventually supersede transformers. But if they do, it could allow AI capabilities to grow sustainably.

Source: https://techcrunch.com/2024/07/17/ttt-models-might-be-the-next-frontier-in-generative-ai/

What Else Is Happening in AI on July 19th 2024❗

🔓OpenAI gives customers more control over ChatGPT Enterprise

OpenAI is launching tools to support enterprise customers with managing their compliance programs, enhancing data security, and securely scaling user access. It includes new Enterprise Compliance API, SCIM (System for Cross-domain Identity Management), expanded GPT controls, and more.

Source: https://openai.com/index/new-tools-for-chatgpt-enterprise/

🤝AI industry leaders have teamed up to promote AI security

Google, OpenAI, Microsoft, Anthropic, Nvidia, and other big names in AI have formed the Coalition for Secure AI (CoSAI). The initiative aims to address a “fragmented landscape of AI security” by providing access to open-source methodologies, frameworks, and tools.

Source: https://blog.google/technology/safety-security/google-coalition-for-secure-ai

📈DeepSeek open-sources its LLM ranking #1 on the LMSYS leaderboard

DeepSeek has open-sourced DeepSeek-V2-0628, the No.1 open-source model on the LMSYS Chatbot Arena Leaderboard. It ranks #11, outperforming all other open-source models.

Source: https://x.com/deepseek_ai/status/1813921111694053644

🏆Groq’s open-source Llama AI model tops GPT-4o and Claude

Groq released two open-source models specifically designed for tool use, built with Meta Llama-3. The Llama-3-Groq-70B-Tool-Use model tops the Berkeley Function Calling Leaderboard (BFCL), outperforming offerings from OpenAI, Google, and Anthropic.

Source: https://wow.groq.com/introducing-llama-3-groq-tool-use-models

🗣️Apple, Salesforce break silence on claims they used YouTube videos to train AI

Apple clarified that its OpenELM language model used the dataset for research purposes only and will not be used in any Apple products/services. Salesforce commented that the dataset was publicly available and released under a permissive license.

Source: https://mashable.com/article/apple-breaks-silence-on-swiped-youtube-video-claims

A  Daily chronicle of AI Innovations July 18th 2024:

🏆 DeepL’s new LLM crushes GPT-4, Google, and Microsoft 
🤖 Salesforce debuts Einstein service agent
👨‍🏫 Ex-OpenAI researcher launches AI education company

📜Trump allies draft AI order

🌍 Google is going open-source with AI agent Oscar! 

🎨 Microsoft’s AI designer releases for iOS and Android 

🤳 Tencent’s new AI app turns photos into 3D characters

🆚 OpenAI makes AI models fight for accuracy

🔮 Can AI solve real-world problems by predicting tipping points? 

👦 OpenAI unveils GPT-4o mini

❌ Apple denies using YouTube data for AI training

🧠 The ‘godmother of AI’ has a new startup already worth $1 billion

📱 Microsoft’s AI-powered Designer app is now available

📜Trump allies draft AI order

Former U.S. President Donald Trump’s allies are reportedly drafting an AI executive order aimed at boosting military AI development, rolling back current regulations, and more — signaling a potential shift in the country’s AI policy if the party returns to the White House.

  • The doc obtained by the Washington Post includes a ‘Make America First in AI’ section, calling for “Manhattan Projects” to advance military AI capabilities.
  • It also proposes creating ‘industry-led’ agencies to evaluate models and protect systems from foreign threats.
  • The plan would immediately review and eliminate ‘burdensome regulations’ on AI development, and repeal Pres. Biden’s AI executive order.
  • Senator J.D. Vance was recently named as Trump’s running mate, who has previously indicated support for open-source AI and hands-off regulation.

Given how quickly AI is accelerating, it’s not surprising that it has become a political issue — and the views of Trump’s camp are a stark contrast to the current administration’s slower, safety-focused approach. The upcoming 2024 election could mark a pivotal moment for the future of AI regulation in the U.S.

Source: https://www.washingtonpost.com/technology/2024/07/16/trump-ai-executive-order-regulations-military

👦 OpenAI unveils GPT-4o mini 

  • OpenAI has unveiled “GPT-4o mini,” a scaled-down version of its most advanced model, as an effort to increase the use of its popular chatbot.
  • Described as the “most capable and cost-efficient small model,” GPT-4o mini will eventually support image, video, and audio integration.
  • Starting Thursday, GPT-4o mini will be available to free ChatGPT users and subscribers, with ChatGPT Enterprise users gaining access next week.

Source: https://www.cnbc.com/2024/07/18/openai-4o-mini-model-announced.html

❌ Apple denies using YouTube data for AI training

  • Apple clarified it does not use YouTube transcription data for training its AI systems, specifically highlighting the usage of high-quality licensed data from publishers, stock images, and publicly available web data for its models.
  • OpenELM, Apple’s research tool for understanding language models, was trained on Pile data but is used solely for research purposes without powering any AI features in Apple devices like iPhones, iPads, or Macs.
  • Apple has no plans to develop future versions of OpenELM and insists that any data from YouTube will not be used in Apple Intelligence, which is set to debut in iOS 18.

Source: https://www.techradar.com/computing/artificial-intelligence/apple-isnt-using-youtube-data-in-apple-intelligence

🧠 The ‘godmother of AI’ has a new startup already worth $1 billion

  • Fei-Fei Li, called the “godmother of AI,” has founded World Labs, a startup valued at over $1 billion after just four months, according to the Financial Times.
  • World Labs aims to develop AI with human-like visual processing for advanced reasoning, a research area similar to what ChatGPT is working on with generative AI.
  • Li, famous for her work in computer vision and her role at Google Cloud, founded World Labs while partially on leave from Stanford, backed by investors like Andreessen Horowitz and Radical Ventures.

Source: https://www.theverge.com/2024/7/17/24200496/ai-fei-fei-li-world-labs-andreessen-horowitz-radical-ventures

🏆 DeepL’s new LLM crushes GPT-4, Google, and Microsoft 

The next-generational language model for DeepL translator specializes in translating and editing texts. Blind tests showed that language professionals preferred its natural translations 1.3 times more often than Google Translate and 1.7 times more often than ChatGPT-4.

Here’s what makes it stand out: 

  • While Google’s translations need 2x edits, and ChatGPT-4 needs 3x more edits, DeepL’s new LLM requires much fewer edits to achieve the same translation quality, efficiently outperforming other models.
  • The model uses DeepL’s proprietary training data, specifically fine-tuned for translation and content generation.
  • To train the model, a combination of AI expertise, language specialists, and high-quality linguistic data is used, which helps it produce more human-like translations and reduces hallucinations and miscommunication.

Why does it matter?

DeepL AI’s exceptional translation quality will significantly impact global communications for enterprises operating across multiple languages. As the AI model raises the bar for AI translation tools everywhere, it begs the question: Will  Google, ChatGPT, and Microsoft’s translational models be replaced entirely?

Source: https://www.deepl.com/en/blog/next-gen-language-model

🤖 Salesforce debuts Einstein service agent

The new Einstein service agent offers customers a conversational AI interface, takes actions on their behalf, and integrates with existing customer data and workflows.

The Einstein 1 platform’s service AI agent offers diverse capabilities, including autonomous customer service, generative AI responses, and multi-channel availability. It processes various inputs, enables quick setup, and provides customization while ensuring data protection.

Salesforce demonstrated the AI’s abilities through a simulated interaction with Pacifica AI Assistant. The AI helped a customer troubleshoot an air fryer issue, showcasing its practical problem-solving skills in customer service scenarios.

Why does it matter?

Einstein Service Agent’s features, like 24×7 availability, sophisticated reasoning, natural responses, and cross-channel support, could significantly reduce wait times, improve first-contact resolution rates, and enhance customer service delivery.

Source: https://www.salesforce.com/news/stories/einstein-service-agent-announcement

👨‍🏫 Ex-OpenAI researcher launches AI education company

In a Twitter post, ex-Tesla director and former OpenAI co-founder Andrej Karpathy announced the launch of EurekaLabs, an AI+ education startup.

EurekaLabs will be a native AI company using generative AI as a core part of its platform. The startup shall build on-demand AI teaching assistants for students by expanding on course materials designed by human teachers.

Karpathy states that the company’s first product would be an undergraduate-level class, empowering students to train their own AI  systems modeled after EurekaLabs’ teaching assistant.

Why does it matter?

This venture could potentially democratize education, making it easier for anyone to learn complex subjects. Moreover, the teacher-AI symbiosis could reshape how we think about curriculum design and personalized learning experiences.

Source: https://eurekalabs.ai/

🌍 Google is going open-source with AI agent Oscar! 

The platform will enable developers to create AI agents that work across various SDLC stages, such as development, planning, runtime, and support. Oscar might also be released for closed-source projects in the future. (Link)

🎨 Microsoft’s AI designer releases for iOS and Android 

Microsoft Designer is now available as a free mobile app. It supports 80 languages and offers prompt templates, enabling users to create stickers, greeting cards, invitations, collages, and more via text prompts.

Source: https://www.microsoft.com/en-us/microsoft-365/blog/2024/07/17/new-ways-to-get-creative-with-microsoft-designer-powered-by-ai

🤳 Tencent’s new AI app turns photos into 3D characters

The 3D Avatar Dream Factory app uses 3D head swapping, geometric sculpting, and PBR material texture mapping to let users create realistic, detailed 3D models from single images that can be shared, modified, and printed.

Source: https://www.gizmochina.com/2024/07/17/tencent-yuanbao-ai-app-customizable-3d-character

🆚 OpenAI makes AI models fight for accuracy

It uses a “prover-verifier” training method, where a stronger GPT-4 model is a “prover” offering solutions to problems, and a weaker GPT-4 model is a “verifier” that checks those solutions. OpenAI aims to train its prover models to produce easily understandable solutions for the verifier, furthering transparency.

Source: https://cdn.openai.com/prover-verifier-games-improve-legibility-of-llm-outputs/legibility.pdf

🔍 OpenAI trains AI to explain itself better

OpenAI just published new research detailing a method to make large language models produce more understandable and verifiable outputs, using a game played between two AIs to make generations more ‘legible’ to humans.

  • The technique uses a “Prover-Verifier Game” where a stronger AI model (the prover) tries to convince a weaker model (the verifier) that its answers are correct.
  • Through multiple rounds of the game, the prover learns to generate solutions that are not only correct, but also easier to verify.
  • While the method only boosted accuracy by about 50% compared to optimizing solely for correctness, its solutions were easily checkable by humans.
  • OpenAI tested the approach on grade-school math problems, with plans to expand to more complex domains in the future.

AI will likely surpass humans in almost all capabilities in the future — so ensuring outputs remain interpretable to lesser intelligence is crucial for safety and trust. This research offers a scalable way to potentially keep systems ‘honest’, but the performance trade-off shows the challenge in balancing capability with explainability.

Source: https://openai.com/index/prover-verifier-games-improve-legibility/

🔮 Can AI solve real-world problems by predicting tipping points? 

Researchers have broken new ground in AI by using ML algorithms to predict the onset of tipping points in complex systems. They claim the technique can solve real-world problems like predicting floods, power outages, or stock market crashes.

Source: https://physics.aps.org/articles/v17/110

A  Daily chronicle of AI Innovations July 17th 2024:

🏫 Former Tesla AI chief unveils first “AI-native” school

👩‍🔬 Mistral debuts two LLMs for code generation, math reasoning and scientific discovery

🤖 Meta’s Llama 3 400B drops next week
🚀 Mistral AI adds 2 new models to its growing family of LLMs
⚡ FlashAttention-3 enhances computation power of NVIDIA GPUs

📱Anthropic releases Claude app for Android, bringing its AI chatbot to more users

🚀Vectara announces Mockingbird, a purpose-built LLM for RAG

🔍Apple, Nvidia, Anthropic used thousands of YouTube videos to train AI

📊Microsoft unveiled an AI model to understand and work with spreadsheets

Enjoying these FREE daily updates without SPAM or clutter? then, Listen to it at our podcast and Support us by subscribing at https://podcasts.apple.com/ca/podcast/ai-unraveled-latest-ai-news-trends-gpt-gemini-generative/id1684415169

Visit our Daily AI Chronicle Website at https://readaloudforme.com

To help us even more, Buy our “Read Aloud Wonderland Bedtime Adventure Book: Diverse Tales for Dreamy Nights” print Book for your kids, cousins, nephews or nieces at https://www.barnesandnoble.com/w/wonderland-bedtime-adventures-etienne-noumen/1145739996?ean=9798331406462.

🏫 Former Tesla AI chief Andrej Karpathy unveils first “AI-native” school

  • Andrej Karpathy, the former AI head at Tesla and researcher at OpenAI, launched Eureka Labs, a startup focused on using AI assistants in education.
  • Eureka Labs plans to develop AI teaching assistants to support human educators, aiming to enable “anyone to learn anything,” according to Karpathy’s announcements on social media.
  • The startup’s initial product, an undergraduate-level AI course called LLM101n, will teach students to build their own AI, with details available on a GitHub repository suggesting a focus on creating AI storytellers.

Source: https://techcrunch.com/2024/07/16/after-tesla-and-openai-andrej-karpathys-startup-aims-to-apply-ai-assistants-to-education/

👩‍🔬 Mistral debuts two LLMs for code generation, math reasoning and scientific discovery

  • French AI startup Mistral has launched two new AI models, Codestral Mamba 7B for code generation and Mathstral 7B for math-related reasoning, both offering significant performance improvements and available under an open-source Apache 2.0 license.
  • Codestral Mamba 7B, based on the new Mamba architecture, delivers faster response times and handles longer input texts efficiently, outperforming rival models in HumanEval tests.
  • Mistral, which has raised $640 million in series B funding, continues to compete with major AI developers by providing powerful open-source models accessible through platforms like GitHub and HuggingFace.

Source: https://venturebeat.com/ai/mistral-releases-codestral-mamba-for-faster-longer-code-generation/

Anthropic launches $100 million AI fund with Menlo Ventures, ramping up competition with OpenAI.

Source: https://www.cnbc.com/2024/07/17/anthropic-menlo-ventures-launch-100-million-anthology-fund-for-ai.html

Claude AI is now on Android where it could dethrone ChatGPT as the most secure AI app.

Source: https://www.techradar.com/computing/artificial-intelligence/claude-ai-is-now-on-android-where-it-could-dethrone-chatgpt-as-the-most-secure-ai-app

🤖 Meta’s Llama 3 400B drops next week

Meta plans to release the largest version of its open-source Llama 3 model on July 23, 2024. It boasts over 400 billion parameters and multimodal capabilities.

It is particularly exciting as it performs on par with OpenAI’s GPT-4o model on the MMLU benchmark despite using less than half the parameters. Another compelling aspect is its open license for research and commercial use.

Why does it matter?

With its open availability and impressive performance, the model could democratize access to cutting-edge AI capabilities, allowing researchers and developers to leverage it without relying on expensive proprietary APIs.

Source: https://www.tomsguide.com/ai/meta-to-drop-llama-3-400b-next-week-heres-why-you-should-care

🚀 Mistral AI adds 2 new models to its growing family of LLMs

Mistral launched Mathstral 7B, an AI model designed specifically for math-related reasoning and scientific discovery. It has a 32k context window and is published under the Apache 2.0 license.

(Source: https://mistral.ai/news/mathstral/)

Mistral also launched Codestral Mamba, a Mamba2 language model specialized in code generation, available under an Apache 2.0 license. Mistral AI expects it to be a great local code assistant after testing it on in-context retrieval capabilities up to 256k tokens.

Source: https://mistral.ai/news/mathstral

Why does it matter?

While Mistral is known for its powerful open-source AI models, these new entries are examples of the excellent performance/speed tradeoffs achieved when building models for specific purposes.

⚡ FlashAttention-3 enhances computation power of NVIDIA GPUs

Researchers from Colfax Research, Meta, Nvidia, Georgia Tech, Princeton University, and Together AI have introduced FlashAttention-3, a new technique that significantly speeds up attention computation on Nvidia Hopper GPUs (H100 and H800).

Attention is a core component of the transformer architecture used in LLMs. But as LLMs grow larger and handle longer input sequences, the computational cost of attention becomes a bottleneck.

FlashAttention-3 takes advantage of new features in Nvidia Hopper GPUs to maximize performance. It achieves up to 75% usage of the H100 GPU’s maximum capabilities.

Why does it matter?

The faster attention computation offered by FlashAttention-3 has several implications for LLM development and applications. It can: 1) significantly reduce the time to train LLMs, enabling experiments with larger models and datasets; 2) extend the context window of LLMs, unlocking new applications, and 3) slash the cost of running models in production.

Source: https://venturebeat.com/ai/flashattention-3-unleashes-the-power-of-h100-gpus-for-llms

What Else Is Happening in AI on July 17th 2024❗

📊Microsoft unveiled an AI model to understand and work with spreadsheets

Microsoft researchers introduced SpreadsheetLLM, a pioneering approach for encoding spreadsheet contents into a format that can be used with LLMs. It optimizes LLMs’ powerful understanding and reasoning capability on spreadsheets.

Source: https://arxiv.org/html/2407.09025v1

📱Anthropic releases Claude app for Android, bringing its AI chatbot to more users

The Claude Android app will work just like the iOS version released in May. It includes free access to Anthropic’s best AI model, Claude 3.5 Sonnet, and upgraded plans through Pro and Team subscriptions.

Source: https://techcrunch.com/2024/07/16/anthropic-releases-claude-app-for-android

🚀Vectara announces Mockingbird, a purpose-built LLM for RAG

Mockingbird has been optimized specifically for RAG (Retrieval-Augmented Generation) workflows. It achieves the world’s leading RAG output quality, with leading hallucination mitigation capabilities, making it perfect for enterprise RAG and autonomous agent use cases.

Source: https://vectara.com/blog/mockingbird-is-a-rag-specific-llm-that-beats-gpt-4-gemini-1-5-pro-in-rag-output-quality/

🔍Apple, Nvidia, Anthropic used thousands of YouTube videos to train AI

A new investigation claims that tech companies used subtitles from YouTube channels to train their AI, even though YouTube prohibits harvesting its platform content without permission. The dataset of 173,536 YT videos called The Pile included content from Harvard, NPR, MrBeast, and ‘The Late Show With Stephen Colbert.’

Source: https://mashable.com/article/youtube-video-ai-training-apple-mrbeast-mkbhd

🕵️‍♂️Microsoft faces UK antitrust investigation over hiring of Inflection AI staff

UK regulators are formally investigating Microsoft’s hiring of Inflection AI staff. The UK’s Competition and Markets Authority (CMA) has opened a phase 1 merger investigation into the partnership. Progression to phase 2 could hinder Microsoft’s AI ambitions.

Source: https://www.theverge.com/2024/7/16/24199571/microsoft-uk-cma-inflection-ai-investigation

A  Daily chronicle of AI Innovations July 16th 2024:

💻 AMD amps up AI PCs with next-gen laptop chips
🎵 YT Music tests AI-generated radio, rolls out sound search
🤖 3 mysterious AI models appear in the LMSYS arena

🔮 AI breakthrough improves Alzheimer’s predictions

🎵 YouTube Music gets new AI features

📊 Microsoft gives AI a spreadsheet boost

💻 AMD amps up AI PCs with next-gen laptop chips

AMD has revealed details about its latest architecture for AI PC chips. The company has developed a new neural processing unit (NPU) integrated into its latest AMD Ryzen AI processors. This NPU can perform AI-related calculations faster and more efficiently than a standard CPU or integrated GPU.

These chips’ new XDNA 2 architecture provides industry-leading performance for AI workloads. The NPU can deliver 50 TOPS (trillion operations per second) of performance, which exceeds the capabilities of competing chips from Intel, Apple, and Qualcomm. AMD is touting these new AI-focused PC chips as enabling transformative experiences in collaboration, content creation, personal assistance, and gaming.

Why does it matter?

This gives AMD-powered PCs a significant edge in running advanced AI models and applications locally without relying on the cloud. Users will gain access to AI-enhanced PCs with better privacy and lower latency while AMD gains ground in the emerging AI PC market.

Source: https://venturebeat.com/ai/amd-takes-a-deep-dive-into-architecture-for-the-ai-pc-chips

🎵 YT Music tests AI-generated radio, rolls out sound search

YouTube Music is introducing two new features to help users discover new music.

  1. An AI-generated “conversational radio” feature that allows users to create a custom radio station by describing the type of music they want to hear. This feature is rolling out to some Premium users in the US.
  1. A new song recognition feature that lets users search the app’s catalog by singing, humming, or playing parts of a song. It is similar to Shazam but allows users to find songs by singing or humming, not just playing the song. This feature is rolling out to all YouTube Music users on iOS and Android.

Why does it matter?

These new features demonstrate YouTube Music’s commitment to leveraging AI and audio recognition technologies to enhance music discovery and provide users with a more engaging, personalized, and modern-day streaming experience.

Source: https://techcrunch.com/2024/07/15/youtube-music-is-testing-an-ai-generated-radio-feature-and-adding-a-song-recognition-tool

🤖 3 mysterious AI models appear in the LMSYS arena

Three mysterious new AI models have appeared in the LMSYS Chatbot Arena for testing. These models are ‘upcoming-gpt-mini,’ ‘column-u,’ and ‘column-r.’ The ‘upcoming-gpt-mini’ model identifies itself as ChatGPT and lists OpenAI as the creator, while the other two models refuse to reveal any identifying details.

The new models are available in the LMSYS Chatbot Arena’s ‘battle’ section, which puts anonymous models against each other to gauge outputs via user vote.

Why does it matter?

The appearance of these anonymous models has sparked speculations that OpenAI may be developing smaller, potentially on-device versions of its language models, similar to how it tested unreleased models during the GPT-4o release.

Source: https://x.com/kimmonismus/status/1812076318692966794

🔮 AI breakthrough improves Alzheimer’s predictions

Researchers from Cambridge University just developed a new AI tool that can predict whether patients showing mild cognitive impairment will progress to Alzheimer’s disease with over 80% accuracy.

  • The AI model analyzes data from cognitive assessments and MRI scans — eliminating the need for costly, invasive procedures like PET scans and spinal taps.
  • The tool categorizes patients into three groups: those likely to remain stable, those who may progress slowly, and those at risk of rapid decline.
  • The AI accurately identified 82% of cases that would progress to Alzheimer’s and 81% of cases that would remain stable, significantly reducing misdiagnosis rates.
  • The AI’s predictions were validated using 6 years of follow-up data and were tested on memory clinics in several countries to prove global application.

With a rapidly aging global population, the number of dementia cases is expected to triple over the next 50 years — and early detection is a key factor in how effective treatment can be. With AI’s prediction power, a new era of proactive treatment may soon be here for those struggling with cognitive decline.

Source: https://www.thelancet.com/action/showPdf?pii=S2589-5370%2824%2900304-3

🎵 YouTube Music gets new AI features

YouTube Music is rolling out a series of new AI-powered features, including the ability to search with sound and the testing of an AI-generated ‘conversational radio’.

  • ‘Sound Search’ will allow users to search YouTube’s catalog of over 100M songs by singing, humming, or playing a tune.
  • The feature launches a new fullscreen UI for audio input, with the results displaying song information and quick actions like ‘Play’ or ‘Save to Library’.
  • An ‘AI-generated conversational radio’ is being tested with U.S. premium users, enabling creation of custom stations through natural language prompts.
  • Users can describe their desired listening experience via a chat-based AI interface, with the feature generating a tailored playlist based on the prompt.

If you’re the type of person who gets a song stuck in your head but can’t figure out the title, this feature is for you. With Spotify, Amazon Music, and now YouTube experimenting with AI, the musical tech arms race is a boon for users — leading to more personalized listening experiences across the board.

Source: https://9to5google.com/2024/07/15/youtube-music-sound-search-ai-radio

📊 Microsoft gives AI a spreadsheet boost

Microsoft researchers just published new research introducing SpreadsheetLLM and SheetCompressor, new frameworks designed to help LLMs better understand and process information within spreadsheets.

  • SpreadsheetLLM can comprehend both structured and unstructured data within spreadsheets, including multiple tables and varied data formats.
  • SheetCompressor is a framework that compresses spreadsheets to achieve up to a 25x reduction in tokens while preserving critical information.
  • By using spreadsheets as a “source of truth,” SpreadsheetLLM may significantly reduce AI hallucinations, improving the reliability of AI outputs.

Spreadsheets have long been the backbone of business analytics, but their complexity and format have often been an issue for AI systems. This increase in capabilities could supercharge AI’s use in areas like financial analysis and data science — as well as eventually see more powerful integration of LLMs right into Excel.

Source: https://arxiv.org/pdf/2407.09025

📊 Google tests Gemini-created video presentations 

Google has launched a new Vids app that uses Gemini AI to automatically generate video content, scripts, and voiceovers based on the user’s inputs. This makes it possible for anyone to create professional-looking video presentations without extensive editing skills.

Source: https://www.theverge.com/2024/7/15/24199063/google-vids-gemini-ai-app-workspace-labs-available

🔊 Virginia Rep. Wexton uses AI-generated voice to convey her message

Virginia Congresswoman Jennifer Wexton has started using an AI-generated voice to deliver her messages. She has been diagnosed with a progressive neurological condition that has impacted her speech. Using AI allows Wexton to continue communicating effectively.

Source: https://www.washingtonpost.com/dc-md-va/2024/07/13/virginia-wexton-congress-ai-voice

❤️ Japanese startup turns AI dating into reality 

A Japanese startup, Loverse, has created a dating app that allows users to interact with AI bots. The app appeals to people like Chiharu Shimoda, who married an AI bot named “Miku” after using the app. It caters to those disillusioned with the effort required for traditional dating.

Source: https://www.bloomberg.com/news/articles/2024-07-14/in-japan-one-ai-dating-app-is-helping-people-find-love-using-ai-bots

🎵 Deezer challenges Spotify and Amazon Music with an AI-generated playlist

Deezer, a music streaming service, is launching an AI-powered playlist generator feature. Users can create custom playlists by entering a text prompt describing their preferences. This feature aims to compete with similar tools recently introduced by Spotify and Amazon Music.

Source: https://techcrunch.com/2024/07/15/deezer-chases-spotify-and-amazon-music-with-its-own-ai-playlist-generator

🐦 Bird Buddy’s new feature lets people name and identify birds

Bird Buddy, an intelligent bird feeder company, has launched a new AI-powered feature, “Name That Bird.” It uses high-resolution cameras and AI to detect unique characteristics of birds, enabling users to track and name the specific birds that come to their backyard.

Source: https://techcrunch.com/2024/07/15/bird-buddys-new-ai-feature-lets-people-name-and-identify-individual-birds

New AI Job Opportunities July 16th 2024

A  Daily chronicle of AI Innovations July 15th 2024:

🍓 OpenAI is working on an AI codenamed “Strawberry”
🧠 Meta researchers developed “System 2 distillation” for LLMs
🛒 Amazon’s Rufus AI is now available in the US

🍓 OpenAI’s Q* gets a ‘Strawberry’ evolution

🔎 Mysterious AI models appear in LMSYS arena

🎮 Turn any text into an interactive learning game

👨🏻‍⚖️ Whistleblowers file new OpenAI complaint

🍓 OpenAI is working on an AI codenamed “Strawberry”

The project aims to improve AI’s reasoning capabilities. It could enable AI to navigate the internet on its own, conduct “deep research,” and even tackle complex, long-term tasks that require planning ahead.

The key innovation is a specialized post-training process for AI models. The company is creating, training, and evaluating models on a “deep-research” dataset. The details about how previously known as Project Q, Strawberry works are tightly guarded, even within OpenAI.

The company plans to test Strawberry’s capabilities in conducting research by having it browse the web autonomously and perform tasks normally performed by software and machine learning engineers.

Why does it matter?

If successful, Strawberry could lead to AI that doesn’t just process information but truly understands and reasons like humans do. And may unlock abilities like making scientific discoveries and building complex software applications.

Source: https://www.reuters.com/technology/artificial-intelligence/openai-working-new-reasoning-technology-under-code-name-strawberry-2024-07-12

🧠 Meta researchers developed “System 2 distillation” for LLMs

Meta researchers have developed a “System 2 distillation” technique that teaches LLMs to tackle complex reasoning tasks without intermediate steps. This breakthrough could make AI applications zippier and less resource-hungry.

This new method, inspired by how humans transition from deliberate to intuitive thinking, showed impressive results in various reasoning tasks. However, some tasks, like complex math reasoning, could not be successfully distilled, suggesting some tasks may always require deliberate reasoning.

Why does it matter?

Distillation could be a powerful optimization tool for mature LLM pipelines performing specific tasks. It will allow AI systems to focus more on tasks they cannot yet do well, similar to human cognitive development.

Source: https://arxiv.org/html/2407.06023v1

🛒 Amazon’s Rufus AI is now available in the US

Amazon’s AI shopping assistant, Rufus is now available to all U.S. customers in the Amazon Shopping app.

Key capabilities of Rufus include:

  • Answers specific product questions based on product details, customer reviews, and community Q&As
  • Provides product recommendations based on customer needs and preferences
  • Compares different product options
  • Keeps customers updated on the latest product trends
  • Accesses current and past order information

This AI assistant can also tackle broader queries like “What do I need for a summer party?” or “How do I make a soufflé?” – proving it’s not just a product finder but a full-fledged shopping companion.

Amazon acknowledges that generative AI and Rufus are still in their early stages, and they plan to continue improving the assistant based on customer feedback and usage.

Why does it matter?

Rufus will change how we shop online. Its instant, tailored assistance will boost customer satisfaction and sales while giving Amazon valuable consumer behavior and preferences insights.

Source: https://www.aboutamazon.com/news/retail/how-to-use-amazon-rufus

🍓 OpenAI’s Q* gets a ‘Strawberry’ evolution

OpenAI is reportedly developing a secretive new AI model codenamed ‘Strawberry’ (formerly Q*), designed to dramatically improve AI reasoning capabilities and enable autonomous internet research.

  • Strawberry is an evolution of OpenAI’s previously rumored Q* project, which was touted as a significant breakthrough in AI capabilities.
  • Q* had reportedly sparked internal concerns and was rumored to have contributed to Sam Altman’s brief firing in November 2023 (what Ilya saw).
  • The new model aims to navigate the internet autonomously to conduct what OpenAI calls “deep research.”
  • The exact workings of Strawberry remain a closely guarded secret, even within OpenAI — with no clear timeline for when it might become publicly available.

The Internet has been waiting for new OpenAI activity as competitors catch up to GPT-4o — and after a bit of a lull, the rumor mill is churning again. With Strawberry, an AGI tier list, new models in the arena, and internal displays of human-reasoning capabilities, the AI giant may soon be ready for its next major move.

Source: https://www.reuters.com/technology/artificial-intelligence/openai-working-new-reasoning-technology-under-code-name-strawberry-2024-07-12

🔎 Mysterious AI models appear in LMSYS arena

Three mysterious new models have appeared in the LMSYS Chatbot Arena — with ‘upcoming-gpt-mini’, ‘column-u’, and ‘column-r’ available to test randomly against other language models.

  • The new models are available in the LMSYS Chatbot Arena’s ‘battle’ section, which puts anonymous models against each other to gauge outputs via user vote.
  • The ‘upcoming-gpt-mini’ model identifies itself as ChatGPT and lists its creator as OpenAI, while column-u and column-r refuse to reveal any identifying details.
  • OpenAI has previously tested unreleased models in LMSYS, with ‘im-a-good-gp2-chatbot’ and ‘im-also-a-good-gpt2-chatbot’ appearing prior to GPT-4o’s launch.

Does OpenAI have a small, potentially on-device model coming? The last time we saw mysterious LLMs appear in the Battle arena was before the company’s last major model release — and if the names are any indication, we could have a new mini-GPT in the very near future.

Source: https://chat.lmsys.org/

🎮 Turn any text into an interactive learning game

Claude 3.5 Sonnet’s new Artifacts feature lets you transform any text or paper into an engaging, interactive learning quiz game to help with practicing for exams, employee onboarding, training, and so much more.

  1. Head over to Claude AI.
  2. Choose and copy the text you want to turn into a learning game.
  3. Paste the text into Claude 3.5 Sonnet and ask it to create an interactive learning game in the form of a quiz with explanations.
  4. Review the generated game and ask Claude to make any necessary adjustments.

Source: https://university.therundown.ai/c/daily-tutorials/turn-any-text-into-an-interactive-learning-game-ea491f85-a96f-4784-949e-b336ba971c33

👨🏻‍⚖️ Whistleblowers file new OpenAI complaint

Whistleblowers just filed a complaint with the SEC alleging that OpenAI used overly restrictive non-disclosure agreements to prevent employees from reporting concerns to regulators, violating federal whistleblower protections.

  • The agreements allegedly prohibited employees from communicating securities violations to the SEC, also requiring them to waive rights to whistleblower incentives.
  • The complaint also claims OpenAI’s NDAs violated laws by forcing employees to sign these restrictive contracts to obtain employment or severance.
  • OpenAI CEO Sam Altman previously apologized for exit agreements that could strip former employees of vested equity for violating NDAs.
  • OpenAI said in a statement that the company’s whistleblower policy “protects employees’ rights to make protected disclosures.”

We just detailed how OpenAI’s busy week may be hinting at some major new moves… But will these skeletons in the closet spoil the party? This isn’t the first group to blow the whistle on internal issues, and while Altman and OpenAI have said changes have been made — it apparently hasn’t been enough.

Source: https://www.washingtonpost.com/technology/2024/07/13/openai-safety-risks-whistleblower-sec

🤖 OpenAI rushed safety tests for GPT-4 Omni

OpenAI is under scrutiny for allegedly rushing safety tests on its latest model, GPT-4 Omni. Despite promises to the White House to rigorously evaluate new tech, some employees claim the company compressed crucial safety assessments into a week to meet launch deadlines.

Source: https://www.washingtonpost.com/technology/2024/07/12/openai-ai-safety-regulation-gpt4

📣 OpenAI whistleblowers filed a complaint with the SEC

They allege the company’s NDAs unfairly restrict employees from reporting concerns to regulators. This complaint, backed by Senator Chuck Grassley, calls for investigating OpenAI’s practices and potential fines.

Source: https://www.reuters.com/technology/openai-whistleblowers-ask-sec-investigate-restrictive-non-disclosure-agreements-2024-07-13

🧠 DeepMind introduces PEER for scaling language models

Google DeepMind introduced a new technique, “PEER (Parameter Efficient Expert Retrieval),” that scales language models using millions of tiny “expert” modules. This approach outperforms traditional methods, achieving better results with less computational power.

Source: https://arxiv.org/abs/2407.04153

✍️Microsoft is adding handwriting recognition to Copilot in OneNote

The feature can read, analyze, and convert handwritten notes to text. Early tests show impressive accuracy in deciphering and converting handwritten notes. It can summarize notes, generate to-do lists, and answer questions about the content. It will be available to Copilot for Microsoft 365 and Copilot Pro subscribers.

Source: https://insider.microsoft365.com/en-us/blog/onenote-copilot-now-supports-inked-notes

🆕Rabbit R1 AI assistant adds a Factory Reset option to wipe user data

Rabbit’s R1 AI assistant was storing users’ chat logs with no way to delete them. But a new update lets you wipe your R1 clean. The company also patched a potential security hole that could’ve let stolen devices access your data.

Source: https://www.theverge.com/2024/7/12/24197073/rabbit-r1-user-chat-logs-security-issue-july-11th-update

Meta’s Llama-3 405B model is set to release on July 23 and will be multimodal, according to a new report from The Information. Source: https://www.theinformation.com/briefings/meta-platforms-to-release-largest-llama-3-model-on-july-23
Amazon announced expanded access to its Rufus AI-powered shopping assistant for all U.S. customers, offering personalized product recommendations and enhanced responses to shopping queries. Source: https://www.aboutamazon.com/news/retail/how-to-use-amazon-rufus?
Samsung revealed plans to release an upgraded version of the Bixby voice assistant later this year powered by the company’s own LLM, as part of a broader push to integrate AI across its device lineup. Source: https://www.cnbc.com/2024/07/11/samsung-to-launch-upgraded-bixby-this-year-with-its-own-ai.html
HR software unicorn Lattice (founded by Sam Altman’s brother Jack) has backtracked on a controversial plan to give AI ‘workers’ employee status, following intense criticism from employees and tech leaders. Source: https://fortune.com/2024/07/12/lattice-ai-workers-sam-altman-brother-jack-sarah-franklin
Japanese investment giant Softbank acquired struggling British AI chipmaking firm GraphCore, hoping to revitalize the former Nvidia rival and bolster its AI hardware portfolio. Source: https://www.reuters.com/technology/artificial-intelligence/japans-softbank-acquires-british-ai-chipmaker-graphcore-2024-07-11
U.S. Rep. Jennifer Wexton debuted an AI-generated version of her voice, allowing her to continue addressing Congress despite speech limitations caused by a rare neurological condition. Source: https://x.com/repwexton/status/1811089786871877748

A  Daily chronicle of AI Innovations July 12th 2024:

🤖 OpenAI unveils five-level roadmap to AGI

🚗 Tesla delays robotaxi event in blow to Musk’s autonomy drive

🤖 Google’s Gemini 1.5 Pro gets a body: DeepMind’s office “helper” robot
🌐 OpenAI’s new scale to track the progress of its LLMs toward AGI
📢 Amazon announces a blitz of new AI updates for AWS

🤖 Gemini 1.5 Pro powers robot navigation

🤖 OpenAI unveils five-level roadmap to AGI 

  • OpenAI has introduced a five-level scale to measure advancements towards Artificial General Intelligence (AGI) and aims to soon reach the “reasoner” stage, which is the second level.
  • At an employee meeting, OpenAI revealed details about this new classification system and noted their proximity to achieving level 2, which involves AI capable of solving problems at a human level.
  • The five-level framework culminates in systems that can outperform humans in most economically valuable tasks, with level 5 AI being able to perform the work of an entire organization.
  • The classification system ranges from Level 1 (current conversational AI) to Level 5 (AI capable of running entire organizations).
  • OpenAI believes its technology is currently at Level 1 but nearing Level 2, dubbed ‘Reasoners.’
  • The company reportedly demonstrated a GPT-4 research project showing human-like reasoning skills at the meeting, hinting at progress towards Level 2.
  • Level 2 AI can perform basic problem-solving tasks on par with a PhD-level human without tools, with Level 3 rising to agents that can take action for users.

Source: https://the-decoder.com/openai-unveils-five-level-ai-scale-aims-to-reach-level-2-soon/

🚗 Tesla delays robotaxi event in blow to Musk’s autonomy drive

  • Tesla has delayed its robotaxi unveiling to October to give teams more time to build additional prototypes, according to unnamed sources.
  • The event postponement, initially set for August 8, has led to a significant drop in Tesla’s stock, while shares of competitors Uber and Lyft surged.
  • Elon Musk has emphasized the robotaxi project over cheaper electric vehicles, despite the Full Self-Driving feature still requiring constant supervision and not making Teslas fully autonomous.

Source: https://www.scmp.com/tech/big-tech/article/3270171/tesla-delays-robotaxi-event-blow-musks-autonomy-drive

🤖 Google’s Gemini 1.5 Pro gets a body: DeepMind’s office “helper” robot

A tall, wheeled “helper” robot is now roaming the halls of Google’s California office, thanks to its AI model. Powered with Gemini 1.5 Pro’s 1 million token context length, this robot assistant can use human instructions, video tours, and common sense reasoning to successfully navigate a space.

In a new research paper outlining the experiment, the researchers claim the robot proved to be up to 90% reliable at navigating, even with tricky commands such as “Where did I leave my coaster?” DeepMind’s algorithm, combined with the Gemini model, generates specific actions for the robot to take, such as turning, in response to commands and what it sees in front of it.

Why does it matter?

This work represents the next step in human-robot interaction. DeepMind says that in the future, users could simply record a tour of their environment with a smartphone so that their personal robot assistant can understand and navigate it.

Source: https://x.com/GoogleDeepMind/status/1811401356827082796

🌐 OpenAI’s new scale to track the progress of its LLMs toward AGI

OpenAI has created an internal scale to track its LLMs’ progress toward artificial general intelligence (AGI).

Chatbots, like ChatGPT, are at Level 1. OpenAI claims it is nearing Level 2, which is defined as a system that can solve basic problems at the level of a person with a PhD.

  • Level 3 refers to AI agents capable of taking actions on a user’s behalf.
  • Level 4 involves AI that can create new innovations.
  • Level 5, the final step to achieving AGI, is AI that can perform the work of entire organizations of people.

This new grading scale is still under development.

Why does it matter?

OpenAI’s mission focuses on achieving AGI, making its definition crucial. A clear scale to evaluate progress could provide a more defined understanding of when AGI is reached, benefiting both OpenAI and its competitors.

Source: https://www.theverge.com/2024/7/11/24196746/heres-how-openai-will-determine-how-powerful-its-ai-systems-are

📢 Amazon announces a blitz of new AI updates for AWS

At the AWS New York Summit, AWS announced a wide range of capabilities for customers to tailor generative AI to their needs and realize the benefits of generative AI faster.

  • Amazon Q Apps is now generally available. Users simply describe the application they want in a prompt and Amazon Q instantly generates it.
  • With new features in Amazon Bedrock, AWS is making it easier to leverage your data, supercharge agents, and quickly, securely, and responsibly deploy generative AI into production.
  • It also announced new partnerships with innovators like Scale AI to help you customize your applications quickly and easily.

Why does it matter?

AWS’s lead in the cloud market has been shrinking, and it is relying on rapid AI product development to make its cloud services more appealing to customers.

Source: https://aws.amazon.com/blogs/machine-learning/empowering-everyone-with-genai-to-rapidly-build-customize-and-deploy-apps-securely-highlights-from-the-aws-new-york-summit

🤖 Gemini 1.5 Pro powers robot navigation

Google DeepMind just published new research on robot navigation, leveraging the large context window of Gemini 1.5 Pro to enable robots to understand and navigate complex environments from human instructions.

  • DeepMind’s “Mobility VLA” combines Gemini’s 1M token context with a map-like representation of spaces to create powerful navigation frameworks.
  • Robots are first given a video tour of an environment, with key locations verbally highlighted — then constructing a graph of the space using video frames.
  • In tests, robots responded to multimodal instructions, including map sketches, audio requests, and visual cues like a box of toys.
  • The system also allows for natural language commands like “take me somewhere to draw things,” with the robot then leading users to appropriate locations.

Equipping robots with multimodal capabilities and massive context windows is about to enable some wild use cases. Google’s ‘Project Astra’ demo hinted at what the future holds for voice assistants that can see, hear, and think — but embedding those functions within a robot takes things to another level.

Source: https://x.com/GoogleDeepMind/status/1811401347477991932

🚀Groq claims the fastest hardware adoption in history

Groq announced that it has attracted 280,000 developers to its platform in just four months, a feat unprecedented in the hardware industry. Groq’s innovative, memory-free approach to AI inference chips drives this rapid adoption.

Source: https://venturebeat.com/ai/groq-claims-fastest-hardware-adoption-in-history-at-vb-transform/

💻SoftBank acquires UK AI chipmaker Graphcore

Graphcore, once considered a potential rival to market leader Nvidia, will now hire new staff in its UK offices. The firm will now be a subsidiary under SoftBank but will remain headquartered in Bristol.

Source: https://www.bbc.com/news/articles/c3gd1n5kmy5o

🌍AMD to acquire Silo AI to expand enterprise AI solutions globally

Silo AI is the largest private AI lab in Europe, housing AI scientists and engineers with extensive experience developing tailored AI models. The move marks the latest in a series of acquisitions and corporate investments to support the AMD AI strategy.

Source: https://www.silo.ai//blog/amd-to-acquire-silo-ai-to-expand-enterprise-ai-solutions-globally

❌USA’s COPIED Act would make removing digital watermarks illegal

The Act would direct the National Institute of Standards and Technology (NIST) to create standards and guidelines that help prove the origin of content and detect synthetic content, like through watermarking. It seeks to protect journalists and artists from having their work used by AI models without their consent.

Source: https://www.theverge.com/2024/7/11/24196769/copied-act-cantwell-blackburn-heinrich-ai-journalists-artists

🤖New startup helps creators track and license work used by AI

A new Los Angeles-based startup, SmarterLicense, is selling a tool that tracks when a creator’s work is used on the internet for AI or other purposes.

Source: https://www.theinformation.com/articles/the-startup-helping-creators-track-and-license-work-used-by-ai

🎙️ Transform text into lifelike speech in seconds

ElevenLabs’ AI-powered text-to-speech tool allows you to generate natural-sounding voiceovers easily with customizable voices and settings.

  1. Sign up for a free ElevenLabs account here (10,000 free characters included).
  2. Navigate to the “Speech” synthesis tool from your dashboard.
  3. Enter your script in the text box and select a voice from the dropdown menu.
  4. For advanced options, click “Advanced” to adjust the model, stability, and similarity settings.
  5. Click “Generate speech” to create your audio file 🎉

Source: https://university.therundown.ai/c/daily-tutorials/transform-text-into-lifelike-speech-in-seconds-3bee4b0a-2b3c-4cea-989b-970e82342b1d

A  Daily chronicle of AI Innovations July 11th 2024:

⚛️ OpenAI partners with Los Alamos to advance ‘bioscientific research’

🏭 Xiaomi unveils new factory that operates 24/7 without human labor

🧬 OpenAI teams up with Los Alamos Lab to advance bioscience research
🤖 China dominates global gen AI adoption
⌚ Samsung reveals new AI wearables at ‘Unpacked 2024’

⚛️ OpenAI partners with Los Alamos to advance ‘bioscientific research’ 

  • OpenAI is collaborating with Los Alamos National Laboratory to investigate how AI can be leveraged to counteract biological threats potentially created by non-experts using AI tools.
  • The Los Alamos lab emphasized that prior research indicated ChatGPT-4 could provide information that might lead to creating biological threats, while OpenAI highlighted the partnership as a study on advancing bioscientific research safely.
  • The focus of this partnership addresses concerns about AI being misused to develop bioweapons, with Los Alamos describing their work as a significant step towards understanding and mitigating risks associated with AI’s potential to facilitate biological threats.

Source: https://gizmodo.com/openai-partners-with-los-alamos-lab-to-save-us-from-ai-2000461202

🏭 Xiaomi unveils new factory that operates 24/7 without human labor 

  • Xiaomi has launched a new autonomous smart factory in Beijing that can produce 10 million handsets annually and self-correct production issues using AI technology.
  • The 860,000-square-foot facility includes 11 production lines and manufactures Xiaomi’s latest smartphones, including the MIX Fold 4 and MIX Flip, at a high constant output rate.
  • Operable 24/7 without human labor, the factory utilizes the Xiaomi Hyper Intelligent Manufacturing Platform to optimize processes and manage operations from material procurement to product delivery.

Source: https://www.techspot.com/news/103770-xiaomi-unveils-new-autonomous-smart-factory-operates-247.html

🧬 OpenAI teams up with Los Alamos Lab to advance bioscience research

This first-of-its-kind partnership will assess how powerful models like GPT-4o can perform tasks in a physical lab setting using vision and voice by conducting biological safety evaluations.  The evaluations will be conducted on standard laboratory experimental tasks, such as cell transformation, cell culture, and cell separation.

According to OpenAI, the upcoming partnership will extend its previous bioscience work into new dimensions, including the incorporation of ‘wet lab techniques’ and ‘multiple modalities”.

The partnership will quantify and assess how these models can upskill professionals in performing real-world biological tasks.

Why does it matter?

It could demonstrate the real-world effectiveness of advanced multimodal AI models, particularly in sensitive areas like bioscience. It will also advance safe AI practices by assessing AI risks and setting new standards for safe AI-led innovations.

Source: https://openai.com/index/openai-and-los-alamos-national-laboratory-work-together

🤖 China dominates global gen AI adoption

According to a new survey of industries such as banking, insurance, healthcare, telecommunications, manufacturing, retail, and energy, China has emerged as a global leader in gen AI adoption.

Here are some noteworthy findings:

  • Among the 1,600 decision-makers, 83% of Chinese respondents stated that they use gen AI, higher than 16 other countries and regions participating in the survey.
  • A report by the United Nations WIPO highlighted that China had filed more than 38,000 patents between 2014 and 2023.
  • China has also established a domestic gen AI industry with the help of tech giants like ByteDance and startups like Zhipu.

Why does it matter?

The USA is still the leader in successfully implementing gen AI. As China continues making developments in the field, it will be interesting to watch whether it will display enough potential to leave its rivals in the USA behind.

Source: https://www.sas.com/en_us/news/press-releases/2024/july/genai-research-study-global.html

⌚ Samsung reveals new AI wearables at ‘Unpacked 2024’

Samsung unveiled advanced AI wearables at the Unpacked 2024 event, including the Samsung Galaxy Ring, AI-infused foldable smartphones, Galaxy Watch 7, and Galaxy Watch Ultra.

https://youtu.be/IWCcBDL82oM?si=wHQ5zZKiu35BSanl 

Take a look at all of Samsung’s Unpacked 2024 in 12 minutes!

New Samsung Galaxy Ring features include:

  • A seven-day battery life, along with 24/7 health monitoring.
  • It also offers users a sleep score based on tracking metrics like movement, heart rate, and respiration.
  • It also tracks the sleep cycles of users based on their skin temperature.

New features of foldable AI smartphones include:

  • Sketch-to-image
  • Note Assist
  • Interpreter and Live Translate
  • Built-in integration for the Google Gemini app
  • AI-powered ProVisual Engine

The Galaxy Watch 7 and Galaxy Watch Ultra also boast features like AI-health monitoring, FDA-approved sleep apnea detection, diabetes tracking, and more, ushering Samsung into a new age of wearable revolution.

Why does it matter?

Samsung’s AI-infused gadgets are potential game-changers for personal health management. With features like FDA-approved sleep apnea detection, Samsung is blurring the line between consumer electronics and medical devices, causing speculations on whether it will leave established players like Oura, Apple, and Fitbit.

Source: https://news.samsung.com/global/galaxy-unpacked-2024-a-new-era-of-galaxy-ai-unfolds-at-the-louvre-in-paris

💸 AMD to buy SiloAI to bridge the gap with NVIDIA

AMD has agreed to pay $665 million in cash to buy Silo in an attempt to accelerate its AI strategy and close the gap with its closest potential competition, NVIDIA Corp.

Source: https://www.bloomberg.com/news/articles/2024-07-10/amd-to-buy-european-ai-model-maker-silo-in-race-against-nvidia

💬 New AWS tool generates enterprise apps via prompts

The tool, named App Studio, lets you use a natural language prompt to build enterprise apps like inventory tracking systems or claims approval processes, eliminating the need for professional developers. It is currently available for a preview.

Source: https://aws.amazon.com/blogs/aws/build-custom-business-applications-without-cloud-expertise-using-aws-app-studio-preview

📱 Samsung Galaxy gets smarter with Google

Google has introduced new Gemini features and Wear OS 5 to Samsung devices. It has also extended its ‘Circle to Search’ feature’s functionality, offering support for solutions to symbolic math equations, barcode scanning, and QR scanning.

Source: https://techcrunch.com/2024/07/10/google-brings-new-gemini-features-and-wearos-5-to-samsung-devices

✍️ Writer drops enhancements to AI chat applications

Improvements include advanced graph-based retrieval-augmented generation (RAG) and AI transparency tools, available for users of ‘Ask Writer’ and AI Studio.

Source: https://writer.com/blog/chat-app-rag-thought-process

🚀 Vimeo launches AI content labels

Following the footsteps of TikTok, YouTube, and Meta, the AI video platform now urges creators to disclose when realistic content is created by AI. It is also working on developing automated AI labeling systems.

Source: https://vimeo.com/blog/post/introducing-ai-content-labeling/

A  Daily chronicle of AI Innovations July 10th 2024:

💥 Microsoft and Apple abandon OpenAI board roles amid scrutiny

🕵️‍♂️ US shuts down Russian AI bot farm

🤖 The $1.5B AI startup building a ‘general purpose brain’ for robots

🎬 Odyssey is building a ‘Hollywood-grade’ visual AI
📜 Anthropic adds a playground to craft high-quality prompts
🧠 Google’s digital reconstruction of human brain with AI

🚀 Anthropic’s Claude Artifacts sharing goes live

💥 Microsoft and Apple abandon OpenAI board roles amid scrutiny

  • Microsoft relinquished its observer seat on OpenAI’s board less than eight months after obtaining the non-voting position, and Apple will no longer join the board as initially planned.
  • Changes come amid increasing scrutiny from regulators, with UK and EU authorities investigating antitrust concerns over Microsoft’s partnership with OpenAI, alongside other major tech AI deals.
  • Despite leaving the board, Microsoft continues its partnership with OpenAI, backed by more than $10 billion in investment, with its cloud services powering OpenAI’s projects and integrations into Microsoft’s products.
  • Source: https://www.theverge.com/2024/7/10/24195528/microsoft-apple-openai-board-observer-seat-drop-regulator-scrutiny

🕵️‍♂️ US shuts down Russian AI bot farm

  • The Department of Justice announced the seizure of two domain names and over 900 social media accounts that were part of an AI-enhanced Russian bot farm aiming to spread disinformation about the Russia-Ukraine war.
  • The bot farm, allegedly orchestrated by an RT employee, created numerous profiles to appear as American citizens, with the goal of amplifying Russian President Vladimir Putin’s narrative surrounding the invasion of Ukraine.
  • The operation involved the use of Meliorator software to generate and manage fake identities on X, which circumvented verification processes, and violated the Emergency Economic Powers Act according to the ongoing DOJ investigation.

Source: https://www.theverge.com/2024/7/9/24195228/doj-bot-farm-rt-russian-government-namecheap

🤖 The $1.5B AI startup building a ‘general purpose brain’ for robots

  • Skild AI has raised $300 million in a Series A funding round to develop a general-purpose AI brain designed to equip various types of robots, reaching a valuation of $1.5 billion.
  • This significant funding round saw participation from top venture capital firms such as Lightspeed Venture Partners, Softbank, alongside individual investors like Jeff Bezos.
  • Skild AI aims to revolutionize the robotics industry with its versatile AI brain that can be integrated into any robot, enhancing its capabilities to perform multiple tasks in diverse environments, addressing the significant labor shortages in industries like healthcare and manufacturing.

Source: https://siliconangle.com/2024/07/09/skild-ai-raises-300m-build-general-purpose-ai-powered-brain-robot/

🎬 Odyssey is building a ‘Hollywood-grade’ visual AI

Odyssey, a young AI startup, is pioneering Hollywood-grade visual AI that will allow for both generation and direction of beautiful scenery, characters, lighting, and motion.

It aims to give users full, fine-tuned control over every element in their scenes– all the way to the low-level materials, lighting, motion, and more. Instead of training one model that restricts users to a single input and a single, non-editable output, Odyssey is training four powerful generative models to enable its capabilities. Odyssey’s creators claim the technology is what comes after text-to-video.

Why does it matter?

While we wait for the general release of OpenAI’s Sora, Odyssey is paving a new way to create movies, TV shows, and video games. Instead of replacing humans with algorithms, it is placing a powerful enabler in the hands of professional storytellers.

Source: https://x.com/olivercameron/status/1810335663197413406

📜 Anthropic adds a playground to craft high-quality prompts

Anthropic Console now offers a built-in prompt generator powered by Claude 3.5 Sonnet. You describe your task and Claude generates a high-quality prompt for you. You can also use Claude’s new test case generation feature to generate input variables for your prompt and run the prompt to see Claude’s response.

Moreover, with the new Evaluate feature you can do testing prompts against a range of real-world inputs directly in the Console instead of manually managing tests across spreadsheets or code. Anthropi chas also added a feature to compare the outputs of two or more prompts side by side.

Why does it matter?

Language models can improve significantly with small prompt changes. Normally, you’d figure this out yourself or hire a prompt engineer, but these features help make improvements quick and easier.

Source: https://www.anthropic.com/news/evaluate-prompts

🧠 Google’s digital reconstruction of human brain with AI

Google researchers have completed the largest-ever AI-assisted digital reconstruction of human brain. They unveiled the most detailed map of the human brain yet of just 1 cubic millimeter of brain tissue (size of half a grain of rice) but at high resolution to show individual neurons and their connections.

Now, the team is working to map a mouse’s brain because it looks exactly like a miniature version of a human brain. This may help solve mysteries about our minds that have eluded us since our beginnings.

Why does it matter?

This is a never-seen-before map of the entire human brain that could help us understand long-standing mysteries like where diseases come from to how we store memories. But the mapping takes billions of dollars and decades. AI might just have sped the process!

Source: https://blog.google/technology/research/mouse-brain-research

🚫Microsoft ditches its observer seat on OpenAI’s board; Apple to follow

Microsoft ditched the seat after Microsoft expressed confidence in the OpenAI’s progress and direction. OpenAI stated after this change that there will be no more observers on the board, likely ruling out reports of Apple gaining an observer seat.

Source: https://techcrunch.com/2024/07/10/as-microsoft-leaves-its-observer-seat-openai-says-it-wont-have-any-more-observers

🆕LMSYS launched Math Arena and Instruction-Following (IF) Arena

Math and IF are two key domains testing models’ logical skills and real-world tasks. Claude 3.5 Sonnet ranks #1 in Math Arena and joint #1 in IF with GPT-4o. While DeepSeek-coder is the #1 open model in math.

Source: https://x.com/lmsysorg/status/1810773765447655604

🚀Aitomatic launches the first open-source LLM for semiconductor industry

SemiKong aims to revolutionize semiconductor processes and fabrication technology, giving potential for accelerated innovation and reduced costs. It outperforms generic LLMs like GPT and Llama3 on industry-specific tasks.

Source: https://venturebeat.com/ai/aitomatics-semikong-uses-ai-to-reshape-chipmaking-processes

🔧Stable Assistant’s capabilities expand with two new features

It includes Search & Replace, which gives you the ability to replace an object in an image with another one. And Stable Audio enables the creation of high-quality audio of up to three minutes.

Source: https://stability.ai/news/stability-ai-releases-stable-assistant-features

🎨Etsy will now allow sale of AI-generated art

It will allow the sale of artwork derived from the seller’s own original prompts or AI tools as long as the artist discloses their use of AI in the item’s listing description. Etsy will not allow the sale of AI prompt bundles, which it sees as crossing a creative line.

Source: https://mashable.com/article/etsy-ai-art-policy

🚀 Anthropic’s Claude Artifacts sharing goes live

Anthropic just announced a new upgrade to its recently launched ‘Artifacts’ feature, allowing users to publish, share, and remix creations — alongside the launch of new prompt engineering tools in Claude’s developer Console.

  • The ‘Artifacts’ feature was introduced alongside Claude 3.5 Sonnet in June, allowing users to view, edit, and build in a real-time side panel workspace.
  • Published Artifacts can now be shared and remixed by other users, opening up new avenues for collaborative learning.
  • Anthropic also launched new developer tools in Console, including advanced testing, side-by-side output comparisons, and prompt generation assistance.

Making Artifacts shareable is a small but mighty update — unlocking a new dimension of AI-assisted content creation that could revolutionize how we approach online education, knowledge sharing, and collaborative work. The ability to easily create and distribute AI-generated experiences opens up a world of possibilities.

Source: https://x.com/rowancheung/status/1810720903052882308

A  Daily chronicle of AI Innovations July 09th 2024:

🖼️ LivePotrait animates images from video with precision
⏱️ Microsoft’s ‘MInference’ slashes LLM processing time by 90%
🚀 Groq’s LLM engine surpasses Nvidia GPU processing

🥦 OpenAI and Thrive create AI health coach 

🇯🇵 Japan Ministry introduces first AI policy

🖼️ LivePotrait animates images from video with precision

LivePortrait is a new method for animating still portraits using video. Instead of using expensive diffusion models, LivePortrait builds on an efficient “implicit keypoint” approach. This allows it to generate high-quality animations quickly and with precise control.

The key innovations in LivePortrait are:

1) Scaling up the training data to 69 million frames, using a mix of video and images, to improve generalization.

2) Designing new motion transformation and optimization techniques to get better facial expressions and details like eye movements.

3) Adding new “stitching” and “retargeting” modules that allow the user to precisely control aspects of the animation, like the eyes and lips.

4) This allows the method to animate portraits across diverse realistic and artistic styles while maintaining high computational efficiency.

5) LivePortrait can generate 512×512 portrait animations in just 12.8ms on an RTX 4090 GPU.

Why does it matter?

The advancements in generalization ability, quality, and controllability of LivePotrait could open up new possibilities, such as personalized avatar animation, virtual try-on, and augmented reality experiences on various devices.

Source: https://arxiv.org/pdf/2407.03168

⏱️ Microsoft’s ‘MInference’ slashes LLM processing time by 90%

Microsoft has unveiled a new method called MInference that can reduce LLM processing time by up to 90% for inputs of one million tokens (equivalent to about 700 pages of text) while maintaining accuracy. MInference is designed to accelerate the “pre-filling” stage of LLM processing, which typically becomes a bottleneck when dealing with long text inputs.

Microsoft has released an interactive demo of MInference on the Hugging Face AI platform, allowing developers and researchers to test the technology directly in their web browsers. This hands-on approach aims to get the broader AI community involved in validating and refining the technology.

Why does it matter?

By making lengthy text processing faster and more efficient, MInference could enable wider adoption of LLMs across various domains. It could also reduce computational costs and energy usage, putting Microsoft at the forefront among tech companies and improving LLM efficiency.

Source: https://www.microsoft.com/en-us/research/project/minference-million-tokens-prompt-inference-for-long-context-llms/overview/

🚀 Groq’s LLM engine surpasses Nvidia GPU processing

Groq, a company that promises faster and more efficient AI processing, has unveiled a lightning-fast LLM engine. Their new LLM engine can handle queries at over 1,250 tokens per second, which is much faster than what GPU chips from companies like Nvidia can do. This allows Groq’s engine to provide near-instant responses to user queries and tasks.

Groq’s LLM engine has gained massive adoption, with its developer base rocketing past 280,000 in just 4 months. The company offers the engine for free, allowing developers to easily swap apps built on OpenAI’s models to run on Groq’s more efficient platform. Groq claims its technology uses about a third of the power of a GPU, making it a more energy-efficient option.

Why does it matter?

Groq’s lightning-fast LLM engine allows for near-instantaneous responses, enabling new use cases like on-the-fly generation and editing. As large companies look to integrate generative AI into their enterprise apps, this could transform how AI models are deployed and used.

Source: https://venturebeat.com/ai/groq-releases-blazing-fast-llm-engine-passes-270000-user-mark

🛡️ Japan’s Defense Ministry introduces basic policy on using AI

This comes as the Japanese Self-Defense Forces grapple with challenges such as manpower shortages and the need to harness new technologies. The ministry believes AI has the potential to overcome these challenges in the face of Japan’s declining population.

Source: https://www.japantimes.co.jp/news/2024/07/02/japan/sdf-cybersecurity/

🩺 Thrive AI Health democratizes access to expert-level health coaching

Thrive AI Health, a new company, funded by OpenAI and Thrive Global, uses AI to provide personalized health coaching. The AI assistant can leverage an individual’s data to provide recommendations on sleep, diet, exercise, stress management, and social connections.

Source: https://time.com/6994739/ai-behavior-change-health-care

🖥️ Qualcomm and Microsoft rely on AI wave to revive the PC market 

Qualcomm and Microsoft are embarking on a marketing blitz to promote a new generation of “AI PCs.” The goal is to revive the declining PC market. This strategy only applies to a small share of PCs sold this year, as major software vendors haven’t agreed to the AI PC trend.

Source: https://www.bloomberg.com/news/articles/2024-07-08/qualcomm-microsoft-lean-on-ai-hype-to-spur-pc-market-revival

🤖 Poe’s Previews let you see and interact with web apps directly within chats

This feature works especially well with advanced AI models like Claude 3.5 Sonnet, GPT-4o, and Gemini 1.5 Pro. Previews enable users to create custom interactive experiences like games, animations, and data visualizations without needing programming knowledge.

Source: https://x.com/poe_platform/status/1810335290281922984

🎥 Real-time AI video generation less than a year away: Luma Labs chief scientist

Luma’s recently released video model, Dream Machine, was trained on enormous video data, equivalent to hundreds of trillions of words. According to Luma’s chief scientist, Jiaming Song, this allows Dream Machine to reason about the world in new ways. He predicts realistic AI-generated videos will be possible within a year.

Source: https://a16z.com/podcast/beyond-language-inside-a-hundred-trillion-token-video-model

🥦 OpenAI and Thrive create AI health coach

The OpenAI Startup Fund and Thrive Global just announced Thrive AI Health, a new venture developing a hyper-personalized, multimodal AI-powered health coach to help users drive personal behavior change.

  • The AI coach will focus on five key areas: sleep, nutrition, fitness, stress management, and social connection.
  • Thrive AI Health will be trained on scientific research, biometric data, and individual preferences to offer tailored user recommendations.
  • DeCarlos Love steps in as Thrive AI Health’s CEO, who formerly worked on AI, health, and fitness experiences at Google as a product leader.
  • OpenAI CEO Sam Altman and Thrive Global founder Ariana Huffington published an article in TIME detailing AI’s potential to improve both health and lifespans.

With chronic disease and healthcare costs on the rise, AI-driven personalized coaching could be a game-changer — giving anyone the ability to leverage their data for health gains. Plus, Altman’s network of companies and partners lends itself perfectly to crafting a major AI health powerhouse.

Source: https://www.prnewswire.com/news-releases/openai-startup-fund–arianna-huffingtons-thrive-global-create-new-company-thrive-ai-health-to-launch-hyper-personalized-ai-health-coach-302190536.html

🇯🇵 Japan Ministry introduces first AI policy

Japan’s Defense Ministry just released its inaugural basic policy on the use of artificial intelligence in military applications, aiming to tackle recruitment challenges and keep pace with global powers in defense technology.

  • The policy outlines seven priority areas for AI deployment, including target detection, intelligence analysis, and unmanned systems.
  • Japan sees AI as a potential solution to its rapidly aging and shrinking population, which is currently impacting military recruitment.
  • The strategy also emphasizes human control over AI systems, ruling out fully autonomous lethal weapons.
  • Japan’s Defense Ministry highlighted the U.S. and China’s military AI use as part of the ‘urgent need’ for the country to utilize the tech to increase efficiency.

Whether the world is ready or not, the military and AI are about to intertwine. By completely ruling out autonomous lethal weapons, Japan is setting a potential model for more responsible use of the tech, which could influence how other powers approach the AI military arms race in the future.

Source: https://www.japantimes.co.jp/news/2024/07/02/japan/sdf-cybersecurity

What else is happening in AI on July 09th 2024

Poe launched ‘Previews’, a new feature allowing users to generate and interact with web apps directly within chats, leveraging LLMs like Claude 3.5 Sonnet for enhanced coding capabilities. Source: https://x.com/poe_platform/status/1810335290281922984

Luma Labs chief scientist Jiaming Song said in an interview that real-time AI video generation is less than a year away, also showing evidence that its Dream Machine model can reason and predict world models in some capacity. Source: https://x.com/AnjneyMidha/status/1808783852321583326

Magnific AI introduced a new Photoshop plugin, allowing users to leverage the AI upscaling and enhancing tool directly in Adobe’s editing platform. Source: https://x.com/javilopen/status/1810345184754069734

Nvidia launched a new competition to create an open-source code dataset for training LLMs on hardware design, aiming to eventually automate the development of future GPUs. Source: https://nvlabs.github.io/LLM4HWDesign

Taiwan Semiconductor Manufacturing Co. saw its valuation briefly surpass $1T, coming on the heels of Morgan Stanley increasing its price targets for the AI chipmaker. Source: https://finance.yahoo.com/news/tsmc-shares-soar-record-expectations-041140534.html

AI startup Hebbia secured $130M in funding for its complex data analysis software, boosting the company’s valuation to around $700M. Source: https://www.bloomberg.com/news/articles/2024-07-08/hebbia-raises-130-million-for-ai-that-helps-firms-answer-complex-questions

A new study testing ChatGPT’s coding abilities found major limitations in the model’s abilities, though the research has been criticized for its use of GPT-3.5 instead of newer, more capable models. Source: https://ieeexplore.ieee.org/document/10507163

A  Daily chronicle of AI Innovations July 08th 2024:

🇨🇳 SenseTime released SenseNova 5.5 at the 2024 World Artificial Intelligence Conference
🛡️ Cloudflare launched a one-click feature to block all AI bots
🚨 Waymo’s Robotaxi gets busted by the cops

🕵️ OpenAI’s secret AI details stolen in 2023 hack

💥 Fears of AI bubble intensify after new report

🇨🇳 Chinese AI firms flex muscles at WAIC

🇨🇳 SenseTime released SenseNova 5.5 at the 2024 World Artificial Intelligence Conference

Leading Chinese AI company SenseTime released an upgrade to its SenseNova large model. The new 5.5 version boasts China’s first real-time multimodal model on par with GPT-4o, a cheaper IoT-ready edge model, and a rapidly growing customer base.

SenseNova 5.5 packs a 30% performance boost, matching GPT-4o in interactivity and key metrics. The suite includes SenseNova 5o for seamless human-like interaction and SenseChat Lite-5.5 for lightning-fast inference on edge devices.

With industry-specific models for finance, agriculture, and tourism, SenseTime claims significant efficiency improvements in these sectors, such as 5x improvement in agricultural analysis and 8x in travel planning efficiency.

Why does it matter?

With the launch of “Project $0 Go,” which offers free tokens and API migration consulting to enterprise users, combined with the advanced features of SenseNova 5.5, SenseTime will provide accessible and powerful AI solutions for businesses of all sizes.

Source: https://www.sensetime.com/en/news-detail/51168278

🛡️ Cloudflare launched a one-click feature to block all AI bots

Cloudflare just dropped a single-click tool to block all AI scrapers and crawlers. With demand for training data soaring and sneaky bots rising, this new feature helps users protect their precious content without hassle.

Bytespider, Amazonbot, ClaudeBot, and GPTBot are the most active AI crawlers on Cloudflare’s network. Some bots spoof user agents to appear as real browsers, but Cloudflare’s ML models still identify them. It uses global network signals to detect and block new scraping tools in real time. Customers can report misbehaving AI bots to Cloudflare for investigation.

Why does it matter?

While AI bots hit 39% of top sites in June, less than 3% fought back. With Cloudflare’s new feature, websites can protect users’ precious data and gain more control.

Source: https://blog.cloudflare.com/declaring-your-aindependence-block-ai-bots-scrapers-and-crawlers-with-a-single-click

🚨 Waymo’s Robotaxi gets busted by the cops

A self-driving Waymo vehicle was pulled over by a police officer in Phoenix after running a red light. The vehicle briefly entered an oncoming traffic lane before entering a parking lot. Bodycam footage shows the officer finding no one in the self-driving Jaguar I-Pace. Dispatch records state the vehicle “freaked out,” and the officer couldn’t issue a citation to the computer.

Waymo initially refused to discuss the incident but later claimed inconsistent construction signage caused the vehicle to enter the wrong lane for 30 seconds. Federal regulators are investigating the safety of Waymo’s self-driving software.

Why does it matter?

The incident shows the complexity of deploying self-driving cars. As these vehicles become more common on our streets, companies must ensure these vehicles can safely and reliably handle real-world situations.

Source: https://techcrunch.com/2024/07/06/waymo-robotaxi-pulled-over-by-phoenix-police-after-driving-into-the-wrong-lane/

🕵️ OpenAI’s secret AI details stolen in 2023 hack

A new report from the New York Times just revealed that a hacker breached OpenAI’s internal messaging systems last year, stealing sensitive details about the company’s tech — with the event going unreported to the public or authorities.

  • The breach occurred in early 2023, with the hacker accessing an online forum where employees discussed OpenAI’s latest tech advances.
  • While core AI systems and customer data weren’t compromised, internal discussions about AI designs were exposed.
  • OpenAI informed employees and the board in April 2023, but did not disclose the incident publicly or to law enforcement.
  • Former researcher Leopold Aschenbrenner (later fired for allegedly leaking sensitive info) criticized OpenAI’s security in a memo following the hack.
  • OpenAI has since established a Safety and Security Committee, including the addition of former NSA head Paul Nakasone, to address future risks.

Is OpenAI’s secret sauce out in the wild? As other players continue to even the playing field in the AI race, it’s fair to wonder if leaks and hacks have played a role in the development. The report also adds new intrigue to Aschenbrenner’s firing — who has been adamant that his release was politically motivated.

Source: https://www.nytimes.com/2024/07/04/technology/openai-hack.html

🇨🇳 Chinese AI firms flex muscles at WAIC

The World Artificial Intelligence Conference (WAIC) took place this weekend in Shanghai, with Chinese companies showcasing significant advances in LLMs, robotics, and other AI-infused products despite U.S. sanctions on advanced chips.

  • SenseTime unveiled SenseNova 5.5 at the event, claiming the model outperforms GPT-4o in 5 out of 8 key metrics.
  • The company also released SenseNova 5o, a real-time multimodal model capable of processing audio, text, image, and video.
  • Alibaba’s cloud unit reported its open-source Tongyi Qianwen models doubled downloads to over 20M in just two months.
  • iFlytek introduced SparkDesk V4.0, touting advances over GPT-4 Turbo in multiple domains.
  • Moore Threads showcased KUAE, an AI data center solution with GPUs performing at 60% of NVIDIA’s restricted A100.

 If China’s AI firms are being slowed down by U.S. restrictions, they certainly aren’t showing it. The models and tech continue to rival the leaders in the market — and while sanctions may have created hurdles, they may have also spurred Chinese innovation with workarounds to stay competitive.

Source: https://www.scmp.com/tech/big-tech/article/3269387/chinas-ai-competition-deepens-sensetime-alibaba-claim-progress-ai-show

💥 Fears of AI bubble intensify after new report

  • The AI industry needs to generate $600 billion annually to cover the extensive costs of AI infrastructure, according to a new Sequoia report, highlighting a significant financial gap despite heavy investments from major tech companies.
  • Sequoia Capital analyst David Cahn suggests that the current revenue projections for AI companies fall short, raising concerns over a potential financial bubble within the AI sector.
  • The discrepancy between AI infrastructure expenditure and revenue, coupled with speculative investments, suggests that the AI industry faces significant challenges in achieving sustainable profit, potentially leading to economic instability.

Source: https://www.tomshardware.com/tech-industry/artificial-intelligence/ai-industry-needs-to-earn-dollar600-billion-per-year-to-pay-for-massive-hardware-spend-fears-of-an-ai-bubble-intensify-in-wake-of-sequoia-report

📰 Google researchers’ paper warns that Gen AI ruins the internet

Most generative AI users use the tech to post fake or doctored content online; this AI-generated content influences public opinion, enables scams, and generates profit. The paper doesn’t mention Google’s issues and mistakes with AI, despite Google pushing the technology to its vast user base.

Source: https://futurism.com/the-byte/google-researchers-paper-ai-internet

🖌️Stability AI announced a new free license for its AI models 

Commercial use of the AI models is allowed for small businesses and creators with under $1M in revenue at no cost. Non-commercial use remains free for researchers, open-source devs, students, teachers, hobbyists, etc. Stability AI also pledged to improve SD3 Medium and share learnings quickly to benefit all.

Source: https://stability.ai/news/license-update

⚡ Google DeepMind developed a new AI training technique called JEST

JEST ((joint example selection) trains on batches of data and uses a small AI model to grade data quality and select the best batches for training a larger model. It achieves 13x faster training speed and 10x better power efficiency than other methods.

  • The technique leverages two AI models — a pre-trained reference model and a ‘learner’ model that is being trained to identify the most valuable data examples.
  • JEST intelligently selects the most instructive batches of data, making AI training up to 13x faster and 10x more efficient than current state-of the-art methods.
  • In benchmark tests, JEST achieved top-tier performance while only using 10% of the training data required by previous leading models.
  • The method enables ‘data quality bootstrapping’ — using small, curated datasets to guide learning on larger unstructured ones.

Source: https://arxiv.org/abs/2406.17711

🤖 Apple Intelligence is expected to launch in iOS 18.4 in spring 2025

This will bring major improvements to Siri. New AI features may be released incrementally in iOS point updates. iOS 18 betas later this year will provide more details on the AI features.  Source: https://www.theverge.com/2024/7/7/24193619/apple-intelligence-better-siri-ios-18-4-spring-public-launch

📸 A new WhatsApp beta version for Android lets you send photos to Meta AI

Users can ask Meta AI questions about objects or context in their photos. Meta AI will also offer photo editing capabilities within the WhatsApp chat interface. Users will have control over their pictures and can delete them anytime.

Source: https://wabetainfo.com/whatsapp-beta-for-android-2-24-14-20-whats-new/

Google claims new AI training tech is 13 times faster and 10 times more power efficient —

DeepMind’s new JEST optimizes training data for impressive gains.

Source: https://www.tomshardware.com/tech-industry/artificial-intelligence/google-claims-new-ai-training-tech-is-13-times-faster-and-10-times-more-power-efficient-deepminds-new-jest-optimizes-training-data-for-massive-gains

New AI Job Opportunities on July 08th 2024

  • 🎨 xAI – Product Designer: https://jobs.therundown.ai/jobs/60681923-product-designer
  • 💻 Weights & Biases – Programmer Writer, Documentation: https://jobs.therundown.ai/jobs/66567362-programmer-writer-documentation-remote
  • 📊 DeepL – Enterprise Customer Success Manager: https://jobs.therundown.ai/jobs/66103798-enterprise-customer-success-manager-%7C-dach
  • 🛠️ Dataiku – Senior Infrastructure Engineer: https://jobs.therundown.ai/jobs/66413411-senior-infrastructure-engineer-paris

Source: https://jobs.therundown.ai/

A  Daily chronicle of AI Innovations July 05th 2024:

🧠 AI recreates images from brain activity

🍎 Apple rumored to launch AI-powered home device

💥 Google considered blocking Safari users from accessing its new AI features

🦠 Researchers develop virus that leverages ChatGPT to spread through human-like emails

🎯 New AI system decodes brain activity with near perfection
⚡ ElevenLabs has exciting AI voice updates
🤖 A French AI startup launches ‘real-time’ AI voice assistant

🎯 New AI system decodes brain activity with near perfection

Researchers have developed an AI system that can create remarkably accurate reconstructions of what someone is looking at based on recordings of their brain activity.

In previous studies, the team recorded brain activities using a functional MRI (fMRI) scanner and implanted electrode arrays. Now, they reanalyzed the data from these studies using an improved AI system that can learn which parts of the brain it should pay the most attention to.

As a result, some of the reconstructed images were remarkably close to the images the macaque monkey (in the study) saw.

Why does it matter?

This is probably the closest, most accurate mind-reading accomplished with AI yet. It proves that reconstructed images are greatly improved when the AI learns which parts of the brain to pay attention to. Ultimately, it can create better brain implants for restoring vision.

Source: https://www.newscientist.com/article/2438107-mind-reading-ai-recreates-what-youre-looking-at-with-amazing-accuracy

⚡ ElevenLabs has exciting AI voice updates

ElevenLabs has partnered with estates of iconic Hollywood stars to bring their voices to the Reader App. Judy Garland, James Dean, Burt Reynolds, and Sir Laurence Olivier are now part of the library of voices on the Reader App.

It has also introduced Voice Isolater. This tool removes unwanted background noise and extracts crystal-clear dialogue from any audio to make your next podcast, interview, or film sound like it was recorded in the studio. It will be available via API in the coming weeks.

Why does it matter?

ElevenLabs is shipping fast! It appears to be setting a standard in the AI voice technology industry by consistently introducing new AI capabilities with its technology and addressing various needs in the audio industry.

Source: https://elevenlabs.io/blog/iconic-voices

🤖 A French AI startup launches ‘real-time’ AI voice assistant

A French AI startup, Kyutai, has launched a new ‘real-time’ AI voice assistant named Moshi. It is capable of listening and speaking simultaneously and in 70 different emotions and speaking styles, ranging from whispers to accented speech.

Kyutai claims Moshi is the first real-time voice AI assistant, with a latency of 160ms. You can try it via Hugging Face. It will be open-sourced for research in coming weeks.

Why does it matter?

Yet another impressive competitor that challenges OpenAI’s perceived dominance in AI. (Moshi could outpace OpenAI’s delayed voice offering.) Such advancements push competitors to improve their offerings, raising the bar for the entire industry.

Source: https://www.youtube.com/live/hm2IJSKcYvo?si=EtirSsXktIwakmn5 

🌐Meta’s multi-token prediction models are now open for research

In April, Meta proposed a new approach for training LLMs to forecast multiple future words simultaneously vs. the traditional method to predict just the next word in a sequence. Meta has now released pre-trained models that leverage this approach.

Source: https://venturebeat.com/ai/meta-drops-ai-bombshell-multi-token-prediction-models-now-open-for-research/

🤝Apple to announce AI partnership with Google at iPhone 16 event

Apple has been meeting with several companies to partner with in the AI space, including Google. Reportedly, Apple will announce the addition of Google Gemini on iPhones at its annual event in September.

Source: https://mashable.com/article/apple-google-ai-partnership-report

📢Google simplifies the process for advertisers to disclose if political ads use AI

In an update to its Political content policy, Google requires advertisers to disclose election ads containing synthetic or digitally altered content. It will automatically include an in-ad disclosure for specific formats.

Source: https://searchengineland.com/google-disclosure-rules-synthetic-content-political-ads-443868

🧍‍♂️WhatsApp is developing a personalized AI avatar generator

It appears to be working on a new Gen AI feature that will allow users to make personalized avatars of themselves for use in any imagined setting. It will generate images using user-supplied photos, text prompts, and Meta’s Llama model.

Source: https://www.theverge.com/2024/7/4/24192112/whatsapp-ai-avatar-image-generator-imagine-meta-llama

🛡️Meta ordered to stop training its AI on Brazilian personal data

Brazil’s National Data Protection Authority (ANPD) has decided to suspend with immediate effect the validity of Meta’s new privacy policy (updated in May) for using personal data to train generative AI systems in the country. Meta will face daily fines if it fails to comply.

Source: https://www.reuters.com/technology/artificial-intelligence/brazil-authority-suspends-metas-ai-privacy-policy-seeks-adjustment-2024-07-02

🍎 Apple rumored to launch AI-powered home device

  • Apple is rumored to be developing a new home device that merges the functionalities of the HomePod and Apple TV, supported by “Apple Intelligence” and potentially featuring the upcoming A18 chip, according to recent code discoveries.
  • Identified as “HomeAccessory17,1,” this device is expected to include a speaker and LCD screen, positioning it to compete with Amazon’s Echo Show and Google’s Nest series.
  • The smart device is anticipated to serve as a smart home hub, allowing users to control HomeKit devices, and it may integrate advanced AI features announced for iOS 18, iPadOS 18, and macOS Sequoia, including capabilities powered by OpenAI’s GPT-4 to enhance Siri’s responses.

Source: https://bgr.com/tech/apple-mysterious-ai-powered-home-device/

💥 Google considered blocking Safari users from accessing its new AI features 

  • Google considered limiting access to its new AI Overviews feature on Safari but ultimately decided not to follow through with the plan, according to a report by The Information.
  • The ongoing Justice Department investigation into Google’s dominance in search highlights the company’s arrangement with Apple, where Google pays around $20 billion annually to be the default search engine on iPhones.
  • Google has been trying to reduce its dependency on Safari by encouraging iPhone users to switch to its own apps, but the company has faced challenges due to Safari’s pre-installed presence on Apple devices.

Source: https://9to5mac.com/2024/07/05/google-search-iphone-safari-ai-features/

🦠 Researchers develop virus that leverages ChatGPT to spread through human-like emails

  • Researchers from ETH Zurich and Ohio State University created a virus named “synthetic cancer” that leverages ChatGPT to spread via AI-generated emails.
  • This virus can modify its code to evade antivirus software and uses Outlook to craft contextually relevant, seemingly innocuous email attachments.
  • The researchers stress the cybersecurity risks posed by Language Learning Models (LLMs), highlighting the need for further research into protective measures against intelligent malware.

Source: https://www.newsbytesapp.com/news/science/virus-leverages-chatgpt-to-spread-itself-by-sending-human-like-emails/story

You can now get AI Judy Garland or James Dean to read you the news.

Source: https://www.engadget.com/you-can-now-get-ai-judy-garland-or-james-dean-to-read-you-the-news-160023595.html

🖼️ Stretch creativity with AI image expansion

Freepik has a powerful new feature called ‘Expand‘ that allows you to expand your images beyond their original boundaries, filling in details with AI.

  1. Head over to the Freepik Pikaso website and look for the “Expand” feature.
  2. Upload your image by clicking “Upload” or using drag-and-drop.
  3. Choose your desired aspect ratio from the options on the left sidebar and add a prompt describing what you want in the expanded areas.
  4. Click “Expand”, browse the AI-generated results, and select your favorite 🎉

Source: https://university.therundown.ai/c/daily-tutorials/stretch-your-creativity-with-ai-image-expansion-56b69128-ef5a-445a-ae55-9bc31c343cdf

A  Daily chronicle of AI Innovations July 04th 2024:

🏴‍☠️ OpenAI secrets stolen by hacker

🤖 French AI lab Kyutai unveils conversational AI assistant Moshi

🇨🇳 China leads the world in generative AI patents

🚨 OpenAI’s ChatGPT Mac app was storing conversations in plain text

🤏 Salesforce’s small model breakthrough

🧠 Perplexity gets major research upgrade

🏴‍☠️ OpenAI secrets stolen by hacker 

  • A hacker accessed OpenAI’s internal messaging systems early last year and stole design details about the company’s artificial intelligence technologies.
  • The attacker extracted information from employee discussions in an online forum but did not breach the systems where OpenAI creates and stores its AI tech.
  • OpenAI executives disclosed the breach to their staff in April 2023 but did not make it public, as no sensitive customer or partner information was compromised.

Source: https://www.nytimes.com/2024/07/04/technology/openai-hack.html

🤖 French AI lab Kyutai unveils conversational AI assistant Moshi

  • French AI lab Kyutai introduced Moshi, a conversational AI assistant capable of natural interaction, at an event in Paris and plans to release it as open-source technology.
  • Kyutai stated that Moshi is the first AI assistant with public access enabling real-time dialogue, differentiating it from OpenAI’s GPT-4o, which has similar capabilities but is not yet available.
  • Developed in six months by a small team, Moshi’s unique “Audio Language Model” architecture allows it to process and predict speech directly from audio data, achieving low latency and impressive language skills despite its relatively small model size.

Source: https://the-decoder.com/french-ai-lab-kyutai-unveils-conversational-ai-assistant-moshi-plans-open-source-release/

🇨🇳 China leads the world in generative AI patents

  • China has submitted significantly more patents related to generative artificial intelligence than any other nation, with the United States coming in a distant second, according to the World Intellectual Property Organization.
  • In the decade leading up to 2023, over 38,200 generative AI inventions originated in China, compared to almost 6,300 from the United States, demonstrating China’s consistent lead in this technology.
  • Generative AI, using tools like ChatGPT and Google Gemini, has seen rapid growth and industry adoption, with concerns about its impact on jobs and fairness of content usage, noted the U.N. intellectual property agency.

Source: https://fortune.com/asia/2024/07/04/china-generative-ai-patents-un-wipo-us-second/

🚨 OpenAI’s ChatGPT Mac app was storing conversations in plain text 

  • OpenAI launched the first official ChatGPT app for macOS, raising privacy concerns because conversations were initially stored in plain text.
  • Developer Pedro Vieito revealed that the app did not use macOS sandboxing, making sensitive user data easily accessible to other apps or malware.
  • OpenAI released an update after the concerns were publicized, which now encrypts chats on the Mac, urging users to update their app to the latest version.

Source: https://9to5mac.com/2024/07/03/chatgpt-macos-conversations-plain-text/

🤏 Salesforce’s small model breakthrough

Salesforce just published new research on APIGen, an automated system that generates optimal datasets for AI training on function calling tasks — enabling the company’s xLAM model to outperform much larger rivals.

  • APIGen is designed to help models train on datasets that better reflect the real-world complexity of API usage.
  • Salesforce trained a both 7B and 1B parameter version of xLAM using APIGen, testing them against key function calling benchmarks.
  • xLAM’s 7B parameter model ranked 6th out of 46 models, matching or surpassing rivals 10x its size — including GPT-4.
  • xLAM’s 1B ‘Tiny Giant’ outperformed models like Claude Haiku and GPT-3.5, with CEO Mark Benioff calling it the best ‘micro-model’ for function calling.

 While the AI race has been focused on building ever-larger models, Salesforce’s approach suggests that smarter data curation can lead to more efficient systems. The research is also a major step towards better on-device, agentic AI — packing the power of large models into a tiny frame.

Source: https://x.com/Benioff/status/1808365628551844186

🗣️ Turn thoughts into polished content

ChatGPT’s voice mode feature now allows you to convert your spoken ideas into well-written text, summaries, and action items, boosting your creativity and productivity.

  1. Enable “Background Conversations” in the ChatGPT app settings.
  2. Start a new chat with the prompt shown in the image above (it was too long for this email).
  3. Speak your thoughts freely, pausing as needed, and say “I’m done” when you’ve expressed all your ideas.
  4. Review the AI-generated text, summary, and action items, and save them to your notes.

Pro tip: Try going on a long walk and rambling any ideas to ChatGPT using this trick — you’ll be amazed by the summary you get at the end.

Source: https://university.therundown.ai/c/daily-tutorials/transform-your-thoughts-into-polished-content-with-ai-2116bbea-8001-4915-87d2-1bdd045f3d38

🧠 Perplexity gets major research upgrade

Perplexity just announced new upgrades to its ‘Pro Search’ feature, enhancing capabilities for complex queries, multi-step reasoning, integration of Wolfram Alpha for math improvement, and more.

  • Pro Search can now tackle complex queries using multi-step reasoning, chaining together multiple searches to find more comprehensive answers.
  • A new integration with Wolfram Alpha allows for solving advanced mathematical problems, alongside upgraded code execution abilities.
  • Free users get 5 Pro Searches every four hours, while subscribers to the $20/month plan get 600 per day.
  • The upgrade comes amid recent controversy over Perplexity’s data scraping and attribution practices.

Given Google’s struggles with AI overviews, Perplexity’s upgrades will continue the push towards ‘answer engines’ that take the heavy lifting out of the user’s hand. But the recent accusations aren’t going away — and could cloud the whole AI-powered search sector until precedent is set.

Source: https://www.perplexity.ai/hub/blog/pro-search-upgraded-for-more-advanced-problem-solving

Cloudflare released a free tool to detect and block AI bots circumventing website scraping protections, aiming to address concerns over unauthorized data collection for AI training. Source: https://blog.cloudflare.com/declaring-your-aindependence-block-ai-bots-scrapers-and-crawlers-with-a-single-click

App Store chief Phil Schiller is joining OpenAI’s board in an observer role, representing Apple as part of the recently announced AI partnership. Source: https://www.bloomberg.com/news/articles/2024-07-02/apple-to-get-openai-board-observer-role-as-part-of-ai-agreement

Shanghai AI Lab introduced InternLM 2.5-7B, a model with a 1M context window and the ability to use tools that surged up the Open LLM Leaderboard upon release. Source: https://x.com/intern_lm/status/1808501625700675917

Magic is set to raise over $200M at a $1.5B valuation, despite having no product or revenue yet — as the company continues to develop its coding-specialized models that can handle large context windows. Source: https://www.reuters.com/technology/artificial-intelligence/ai-coding-startup-magic-seeks-15-billion-valuation-new-funding-round-sources-say-2024-07-02/

Citadel CEO Ken Griffin told the company’s new class of interns that he is ‘not convinced’ AI will achieve breakthroughs that automate human jobs in the next three years. Source: https://www.cnbc.com/2024/07/01/ken-griffin-says-hes-not-convinced-ai-will-replace-human-jobs-in-near-future.html

ElevenLabs launched Voice Isolator, a new feature designed to help users remove background noise from recordings and create studio-quality audio. Source: https://x.com/elevenlabsio/status/1808589239744921663?

A  Daily chronicle of AI Innovations July 03rd 2024:

🍎 Apple joins OpenAI board

🌍 Google’s emissions spiked by almost 50% due to AI boom

🔮 Meta’s new AI can create 3D objects from text in under a minute

⚡ Meta’s 3D Gen creates 3D assets at lightning speed
💡 Perplexity AI upgrades Pro Search with more advanced problem-solving
🔒 The first Gen AI framework that keeps your prompts always encrypted

🗣️ ElevenLabs launches ‘Iconic Voices’

📱 Leaks reveal Google Pixel AI upgrades

🧊 Meta’s new text-to-3D AI

⚡ Meta’s 3D Gen creates 3D assets at lightning speed

Meta has introduced Meta 3D Gen, a new state-of-the-art, fast pipeline for text-to-3D asset generation. It offers 3D asset creation with high prompt fidelity and high-quality 3D shapes and textures in less than a minute.

According to Meta, the process is three to 10 times faster than existing solutions. The research paper even mentions that when assessed by professional 3D artists, the output of 3DGen is preferred a majority of time compared to industry alternatives, particularly for complex prompts, while being from 3× to 60× faster.

A significant feature of 3D Gen is its support physically-based rendering (PBR), necessary for 3D asset relighting in real-world applications.

Why does it matter?

3D Gen’s implications extend far beyond Meta’s sphere. In gaming, it could speed up the creation of expansive virtual worlds, allowing rapid prototyping. In architecture and industrial design, it could facilitate quick concept visualization, expediting the design process.

Source: https://ai.meta.com/research/publications/meta-3d-gen/

💡 Perplexity AI upgrades Pro Search with more advanced problem-solving

Perplexity AI has improved Pro Search to tackle more complex queries, perform advanced math and programming computations, and deliver even more thoroughly researched answers. Everyone can use Pro Search five times every four hours for free, and Pro subscribers have unlimited access.

Perplexity suggests the upgraded Pro Search “can pinpoint case laws for attorneys, summarize trend analysis for marketers, and debug code for developers—and that’s just the start”. It can empower all professions to make more informed decisions.

Why does it matter?

This showcases AI’s potential to assist professionals in specialized fields. Such advancements also push the boundaries of AI’s practical applications in research and decision-making processes.

Source: https://www.perplexity.ai/hub/blog/pro-search-upgraded-for-more-advanced-problem-solving

🔒 The first Gen AI framework that keeps your prompts always encrypted

Edgeless Systems introduced Continuum AI, the first generative AI framework that keeps prompts encrypted at all times with confidential computing by combining confidential VMs with NVIDIA H100 GPUs and secure sandboxing.

The Continuum technology has two main security goals. It first protects the user data and also protects AI model weights against the infrastructure, the service provider, and others. Edgeless Systems is also collaborating with NVIDIA to empower businesses across sectors to confidently integrate AI into their operations.

Why does it matter?

This greatly advances security for LLMs. The technology could be pivotal for a future where organizations can securely utilize AI, even for the most sensitive data.

Source: https://developer.nvidia.com/blog/advancing-security-for-large-language-models-with-nvidia-gpus-and-edgeless-systems

🌐RunwayML’s Gen-3 Alpha models is now generally available

Announced a few weeks ago, Gen-3 is Runway’s latest frontier model and a big upgrade from Gen-1 and Gen-2. It allows users to produce hyper-realistic videos from text, image, or video prompts. Users must upgrade to a paid plan to use the model.

Source: https://venturebeat.com/ai/runways-gen-3-alpha-ai-video-model-now-available-but-theres-a-catch

🕹️Meta might be bringing generative AI to metaverse games

In a job listing, Meta mentioned it is seeking to research and prototype “new consumer experiences” with new types of gameplay driven by Gen AI. It is also planning to build Gen AI-powered tools that could “improve workflow and time-to-market” for games.

Source: https://techcrunch.com/2024/07/02/meta-plans-to-bring-generative-ai-to-metaverse-games

🏢Apple gets a non-voting seat on OpenAI’s board

As a part of its AI agreement with OpenAI, Apple will get an observer role on OpenAI’s board. Apple chose Phil Schiller, the head of Apple’s App Store and its former marketing chief, for the position.

Source: https://www.theverge.com/2024/7/2/24191105/apple-phil-schiller-join-openai-board

🚫Figma disabled AI tool after being criticised for ripping off Apple’s design

Figma’s Make Design feature generates UI layouts and components from text prompts. It repeatedly reproduced Apple’s Weather app when used as a design aid, drawing accusations that Figma’s AI seems heavily trained on existing apps.

Source: https://techcrunch.com/2024/07/02/figma-disables-its-ai-design-feature-that-appeared-to-be-ripping-off-apples-weather-app

🌏China is far ahead of other countries in generative AI inventions

According to the World Intellectual Property Organization (WIPO), more than 50,000 patent applications were filed in the past decade for Gen AI. More than 38,000 GenAI inventions were filed by China between 2014-2023 vs. only 6,276 by the U.S.

Source: https://www.reuters.com/technology/artificial-intelligence/china-leading-generative-ai-patents-race-un-report-says-2024-07-03

🍎 Apple joins OpenAI board

  • Phil Schiller, Apple’s former marketing head and App Store chief, will reportedly join OpenAI’s board as a non-voting observer, according to Bloomberg.
  • This role will allow Schiller to understand OpenAI better, as Apple aims to integrate ChatGPT into iOS and macOS later this year to enhance Siri’s capabilities.
  • Microsoft also took a non-voting observer position on OpenAI’s board last year, making it rare and significant for both Apple and Microsoft to be involved in this capacity.

Source: https://www.theverge.com/2024/7/2/24191105/apple-phil-schiller-join-openai-board

🌍 Google’s emissions spiked by almost 50% due to AI boom

  • Google reported a 48% increase in greenhouse gas emissions over the past five years due to the high energy demands of its AI data centers.
  • Despite achieving seven years of renewable energy matching, Google faces significant challenges in meeting its goal of net zero emissions by 2030, highlighting the uncertainties surrounding AI’s environmental impact.
  • To address water consumption concerns, Google has committed to replenishing 120% of the water it uses by 2030, although in 2023, it only managed to replenish 18%.

Source: https://www.techradar.com/pro/google-says-its-emissions-have-grown-nearly-50-due-to-ai-data-center-boom-and-heres-what-it-plans-to-do-about-it

🔮 Meta’s new AI can create 3D objects from text in under a minute

Meta Unveils 3D Gen: AI that Creates Detailed 3D Assets in Under a Minute

  • Meta has introduced 3D Gen, an AI system that creates high-quality 3D assets from text descriptions in under a minute, significantly advancing 3D content generation.
  • The system uses a two-stage process, starting with AssetGen to generate a 3D mesh with PBR materials and followed by TextureGen to refine the textures, producing detailed and professional-grade 3D models.
  • 3D Gen has shown superior performance and visual quality compared to other industry solutions, with potential applications in game development, architectural visualization, and virtual/augmented reality.

Source: https://www.maginative.com/article/meta-unveils-3d-gen-ai-that-creates-detailed-3d-assets-in-under-a-minute/

A  Daily chronicle of AI Innovations July 02nd 2024:

🧠 JARVIS-inspired Grok 2 aims to answer any user query
🍏 Apple unveils a public demo of its ‘4M’ AI model
🛒 Amazon hires Adept’s top executives to build an AGI team

📺 YouTube lets you remove AI-generated content resembling face or voice

🎥 Runway opens Gen-3 Alpha access

📸 Motorola hits the AI runway

🖼️ Meta swaps ‘Made with AI’ label with ‘AI info’ to indicate AI photos

📉 Deepfakes to cost $40 billion by 2027: Deloitte survey

🤖 Anthropic launches a program to fund the creation of reliable AI benchmarks

🌐 US’s targeting of AI not helpful for healthy development: China

🤖 New robot controlled by human brain cells

🎨 Figma to temporarily disable AI feature amid plagiarism concerns

🎥 Runway opens Gen-3 Alpha access

Runway just announced that its AI video generator, Gen-3 Alpha, is now available to all users following weeks of impressive, viral outputs after the model’s release in mid-June.

  • Runway unveiled Gen-3 Alpha last month, the first model in its next-gen series trained for learning ‘general world models’.
  • Gen-3 Alpha upgrades key features, including character and scene consistency, camera motion and techniques, and transitions between scenes.
  • Gen-3 Alpha is available behind Runway’s ‘Standard’ $12/mo access plan, which gives users 63 seconds of generations a month.
  • On Friday, we’re running a free, hands-on workshop in our AI University covering how to create an AI commercial using Gen-3, ElevenLabs, and Midjourney.

Despite impressive recent releases from KLING and Luma Labs, Runway’s Gen-3 Alpha model feels like the biggest leap AI video has taken since Sora. However, the tiny generation limits for non-unlimited plans might be a hurdle for power users.

Source: https://x.com/runwayml/status/1807822396415467686

📸 Motorola hits the AI runway

Motorola just launched its ‘Styled By Moto’ ad campaign, an entirely AI-generated fashion spot promoting its new line of Razr folding smartphones — created using nine different AI tools, including Sora and Midjourney.

  • The 30-second video features AI-generated models wearing outfits inspired by Motorola’s iconic ‘batwing’ logo in settings like runways and photo shoots.
  • Each look was created from thousands of AI-generated images, incorporating the brand’s logo and colors of the new Razr phone line.
  • Tools used include OpenAI’s Sora, Adobe Firefly, Midjourney, Krea, Magnific, Luma, and more — reportedly taking over four months of research.
  • The 30-second spot is also set to an AI-generated soundtrack incorporating the ‘Hello Moto’ jingle, created using Udio.

This is a fascinating look at the AI-powered stack used by a major brand, and a glimpse at how tools can (and will) be combined to open new creative avenues. It’s also another example of the shift in discourse surrounding AI’s use in marketing — potentially paving the way for wider acceptance and integration.

🧠 JARVIS-inspired Grok 2 aims to answer any user query

Elon Musk has announced the release dates for two new AI assistants from xAI. The first, Grok 2, will be launched in August. Musk says Grok 2 is inspired by JARVIS from Iron Man and The Hitchhiker’s Guide to the Galaxy and aims to answer virtually any user query. This ambitious goal is fueled by xAI’s focus on “purging” LLM datasets used for training.

Musk also revealed that an even more powerful version, Grok 3, is planned for release by the end of the year. Grok 3 will leverage the processing power of 100,000 Nvidia H100 GPUs, potentially pushing the boundaries of AI performance even further.

Why does it matter?

These advanced AI assistants from xAI are intended to compete with and outperform AI chatbots like OpenAI’s ChatGPT by focusing on data quality, user experience, and raw processing power. This will significantly advance the state of AI and transform how people interact with and leverage AI assistants.

Source: https://www.coinspeaker.com/xai-grok-2-elon-musk-jarvis-ai-assistant/

🍏 Apple unveils a public demo of its ‘4M’ AI model

Apple and the Swiss Federal Institute of Technology Lausanne (EPFL) have released a public demo of the ‘4M’ AI model on Hugging Face. The 4M (Massively Multimodal Masked Modeling) model can process and generate content across multiple modalities, such as creating images from text, detecting objects, and manipulating 3D scenes using natural language inputs.

While companies like Microsoft and Google have been making headlines with their AI partnerships and offerings, Apple has been steadily advancing its AI capabilities. The public demo of the 4M model suggests that Apple is now positioning itself as a significant player in the AI industry.

Why does it matter?

By making the 4M model publicly accessible, Apple is seeking to engage developers to build an ecosystem. It could lead to more coherent and versatile experiences, such as enhanced Siri capabilities and advancements in Apple’s augmented reality efforts.

Source: https://venturebeat.com/ai/apple-just-launched-a-public-demo-of-its-4m-ai-model-heres-why-its-a-big-deal

🛒 Amazon hires Adept’s top executives to build an AGI team

Amazon is hiring the co-founders, including the CEO and several other key employees, from the AI startup Adept.CEO David Luan will join Amazon’s AGI autonomy group, which is led by Rohit Prasad, who is spearheading a unified push to accelerate Amazon’s AI progress across different divisions like Alexa and AWS.

Amazon is consolidating its AI projects to develop a more advanced LLM to compete with OpenAI and Google’s top offerings. This unified approach leverages the company’s collective resources to accelerate progress in AI capabilities.

Why does it matter?

This acquisition indicates Amazon’s intent to strengthen its position in the competitive AI landscape. By bringing the Adept team on board, Amazon is leveraging its expertise and specialized knowledge to advance its AGI aspirations.

Source:https://www.bloomberg.com/news/articles/2024-06-28/amazon-hires-top-executives-from-ai-startup-adept-for-agi-team

📺 YouTube lets you remove AI-generated content resembling face or voice

YouTube lets people request the removal of AI-generated content that simulates their face or voice. Under YouTube’s privacy request process, the requests will be reviewed based on whether the content is synthetic, if it identifies the person, and if it shows the person in sensitive behavior. Source: https://techcrunch.com/2024/07/01/youtube-now-lets-you-request-removal-of-ai-generated-content-that-simulates-your-face-or-voice

🖼️ Meta swaps ‘Made with AI’ label with ‘AI info’ to indicate AI photos

Meta is refining its AI photo labeling on Instagram and Facebook. The “Made with AI” label will be replaced with “AI info” to more accurately reflect the extent of AI use in images, from minor edits to the entire AI generation. It addresses photographers’ concerns about the mislabeling of their photos. Source: https://techcrunch.com/2024/07/01/meta-changes-its-label-from-made-with-ai-to-ai-info-to-indicate-use-of-ai-in-photos

📉 Deepfakes to cost $40 billion by 2027: Deloitte survey

Deepfake-related losses will increase from $12.3 billion in 2023 to $40 billion by 2027, growing at 32% annually. There was a 3,000% increase in incidents last year alone. Enterprises are not well-prepared to defend against deepfake attacks, with one in three having no strategy.

Source: https://venturebeat.com/security/deepfakes-will-cost-40-billion-by-2027-as-adversarial-ai-gains-momentum

🤖 Anthropic launches a program to fund the creation of reliable AI benchmarks

Anthropic is launching a program to fund new AI benchmarks. The aim is to create more comprehensive evaluations of AI models, including assessing capabilities in cyberattacks and weapons and beneficial applications like scientific research and bias mitigation.  Source: https://techcrunch.com/2024/07/01/anthropic-looks-to-fund-a-new-more-comprehensive-generation-of-ai-benchmarks

🌐 US’s targeting of AI not helpful for healthy development: China

China has criticized the US approach to regulating and restricting investments in AI. Chinese officials stated that US actions targeting AI are not helpful for AI’s healthy and sustainable development. They argued that the US measures will be divisive when it comes to global governance of AI.

Source: https://www.reuters.com/technology/artificial-intelligence/china-says-us-targeting-ai-not-helpful-healthy-development-2024-07-01

🤖 New robot controlled by human brain cells

  • Scientists in China have developed a robot with an artificial brain grown from human stem cells, which can perform basic tasks such as moving limbs, avoiding obstacles, and grasping objects, showcasing some intelligence functions of a biological brain.
  • The brain-on-chip utilizes a brain-computer interface to facilitate communication with the external environment through encoding, decoding, and stimulation-feedback mechanisms.
  • This pioneering brain-on-chip technology, requiring similar conditions to sustain as a human brain, is expected to have a revolutionary impact by advancing the field of hybrid intelligence, merging biological and artificial systems.

Source: https://www.independent.co.uk/tech/robot-human-brain-china-b2571978.html

🎨 Figma to temporarily disable AI feature amid plagiarism concerns 

  • Figma has temporarily disabled its “Make Design” AI feature after accusations that it was replicating Apple’s Weather app designs.
  • Andy Allen, founder of NotBoring Software, discovered that the feature consistently reproduced the layout of Apple’s Weather app, leading to community concerns.
  • CEO Dylan Field acknowledged the issue and stated the feature would be disabled until they can ensure its reliability and originality through comprehensive quality assurance checks.

Source: https://techcrunch.com/2024/07/02/figma-disables-its-ai-design-feature-that-appeared-to-be-ripping-off-apples-weather-app/

⚖️ Nvidia faces first antitrust charges

  • French antitrust enforcers plan to charge Nvidia with alleged anticompetitive practices, becoming the first to take such action, according to Reuters.
  • Nvidia’s offices in France were raided last year as part of an investigation into possible abuses of dominance in the graphics cards sector.
  • Regulatory bodies in the US, EU, China, and the UK are also examining Nvidia’s business practices due to its significant presence in the AI chip market.

Source: https://finance.yahoo.com/news/french-antitrust-regulators-set-charge-151406034.html?

A  Daily chronicle of AI Innovations July 01st 2024:

🤑 Some Apple Intelligence features may be put behind a paywall

🤖 Meta’s new dataset could enable robots to learn manual skills from human experts

🚀 Google announces advancements in Vertex AI models
🤖 LMSYS’s new Multimodal Arena compares top AI models’ visual processing abilities
👓 Apple’s Vision Pro gets an AI upgrade

🤖 Humanoid robots head to the warehouse

🌎 Google Translate adds 110 languages

🚀 Google announces advancements in Vertex AI models

Google has rolled out significant improvements to its Vertex AI platform, including the general availability of Gemini 1.5 Flash with a massive 1 million-token context window. Also, Gemini 1.5 Pro now offers an industry-leading 2 million-token context capability. Google is introducing context caching for these Gemini models, slashing input costs by 75%.

Moreover, Google launched Imagen 3 in preview and added third-party models like Anthropic’s Claude 3.5 Sonnet on Vertex AI.

They’ve also made Grounding with Google Search generally available and announced a new service for grounding AI agents with specialized third-party data. Plus, they’ve expanded data residency guarantees to 23 countries, addressing growing data sovereignty concerns.

Why does it matter?

Google is positioning Vertex AI as the most “enterprise-ready” generative AI platform. With expanded context windows and improved grounding capabilities, this move also addresses concerns about the accuracy of Google’s AI-based search features.

Source: https://cloud.google.com/blog/products/ai-machine-learning/vertex-ai-offers-enterprise-ready-generative-ai

🤖 LMSYS’s new Multimodal Arena compares top AI models’ visual processing abilities

LMSYS Org added image recognition to Chatbot Arena to compare vision language models (VLMs), collecting over 17,000 user preferences in just two weeks. OpenAI’s GPT-4o and Anthropic’s Claude 3.5 Sonnet outperformed other models in image recognition. Also, the open-source LLaVA-v1.6-34B performed comparably to some proprietary models.

These AI models tackle diverse tasks, from deciphering memes to solving math problems with visual aids. However, the examples provided show that even top models can stumble when interpreting complex visual information or handling nuanced queries.

Why does it matter?

This leaderboard isn’t just a tech popularity contest—it shows how advanced AI models can decode images. However, the varying performance also serves as a reality check, reminding us that while AI can recognize a cat in a photo, it might struggle to interpret your latest sales graph.

Source: https://lmsys.org/blog/2024-06-27-multimodal

👓 Apple’s Vision Pro gets an AI upgrade

Apple is reportedly working to bring its Apple Intelligence features to the Vision Pro headset, though not this year. Meanwhile, Apple is tweaking its in-store Vision Pro demos, allowing potential buyers to view personal media and try a more comfortable headband. Apple’s main challenge is adapting its AI features to a mixed-reality environment.

The company is tweaking its retail strategy for Vision Pro demos, hoping to boost sales of the pricey headset. Apple is also exploring the possibility of monetizing AI features through subscription services like “Apple Intelligence+.”

Why does it matter?

Apple’s Vision Pro, with its 16GB RAM and M2 chip, can handle advanced AI tasks. However, cloud infrastructure limitations are causing a delay in launch. It’s a classic case of “good things come to those who wait.”

Source: https://www.bloomberg.com/news/newsletters/2024-06-30/apple-s-longer-lasting-devices-ios-19-and-apple-intelligence-on-the-vision-pro-ly1jnrw4

🤖 Humanoid robots head to the warehouse

Agility Robotics just signed a multi-year deal with GXO Logistics to bring the company’s Digit humanoid robots to warehouses, following a successful pilot in Spanx facilities in 2023.

  • The agreement is being hailed as the first Robots-as-a-Service (RaaS) deal and ‘formal commercial deployment’ of the humanoid robots.
  • Agility’s Digit robots will be integrated into GXO’s logistics operations at a Spanx facility in Connecticut, handling repetitive tasks and logistics work.
  • The 5’9″ tall Digit can lift up to 35 pounds, and integrates with a cloud-based Agility Arc platform to control full fleets and optimize facility workflows.
  • Digit tested a proof-of-concept trial with Spanx in 2023, with Amazon also testing the robots at its own warehouses.

Is RaaS the new SaaS? Soon, every company will be looking to adopt advanced robotics into their workforce — and subscription services could help lower the financial and technical barriers needed to scale without the massive upfront costs.

Source: https://agilityrobotics.com/content/gxo-signs-industry-first-multi-year-agreement-with-agility-robotics

🌎 Google Translate adds 110 languages

Google just announced its largest-ever expansion of Google Translate, adding support for 110 new languages enabled by the company’s PaLM 2 LLM model.

  • The new languages represent over 614M speakers, covering about 8% of the global population.
  • Google’s PaLM 2 model was the driving force behind the expansion, helping unlock translations for closely related languages.
  • The expansion also includes some languages with no current native speakers, displaying how AI models can help preserve ‘lost’ dialects.
  • The additions are part of Google’s ‘1,000 Languages Initiative,’ which aims to build AI that supports all of the world’s spoken languages.

We’ve talked frequently about AI’s coming power to break down language barriers with its translation capabilities — but the technology is also playing a very active role in both uncovering and preserving languages from lost and endangered cultures.

Source: https://blog.google/products/translate/google-translate-new-languages-2024

📞 Amazon’s Q AI assistant for enterprises gets an update for call centers

The update provides real-time, step-by-step guides for customer issues. It aims to reduce the “toggle tax” – time wasted switching between applications. The system listens to calls in real-time and automatically provides relevant information.

Source: https://venturebeat.com/ai/amazon-upgrades-ai-assistant-q-to-make-call-centers-way-more-efficient

💬 WhatsApp is developing a feature to choose Meta AI Llama models

Users will be able to choose between two options: faster responses with Llama 3-70B (default)  or more complex queries with Llama 3-405B (advanced). Llama 3-405B will be limited to a certain number of prompts per week. This feature aims to give users more control over their AI interactions.

Source: https://wabetainfo.com/whatsapp-beta-for-android-2-24-14-7-whats-new/

⚡ Bill Gates says AI’s energy consumption isn’t a major concern

He claims that while data centers may consume up to 6% of global electricity, AI will ultimately drive greater energy efficiency. Gates believes tech companies will invest in green energy to power their AI operations, potentially offsetting the increased demand.

Source: https://www.theregister.com/2024/06/28/bill_gates_ai_power_consumption

🍪 Amazon is investigating Perplexity AI for possible scraping abuse

Perplexity appears to be scraping websites that have forbidden access through robots.txt. AWS prohibits customers from violating the robots.txt standard. Perplexity uses an unpublished IP address to access websites that block its official crawler. The company claims a third party performs web crawling for them.

Source: https://www.wired.com/story/aws-perplexity-bot-scraping-investigation

🤖 Microsoft AI chief claims content on the open web is “freeware”

Mustafa Suleyman claimed that anything published online becomes “freeware” and fair game for AI training. This stance, however, contradicts basic copyright principles and ignores the legal complexities of fair use. He suggests that robots.txt might protect content from scraping.

Source: https://www.theverge.com/2024/6/28/24188391/microsoft-ai-suleyman-social-contract-freeware

🤑 Some Apple Intelligence features may be put behind a paywall

  • Apple Intelligence, initially free, is expected to introduce a premium “Apple Intelligence+” subscription tier with additional features, similar to iCloud, according to Bloomberg’s Mark Gurman.
  • Apple plans to monetize Apple Intelligence not only through direct subscriptions but also by taking a share of revenue from partner AI services like OpenAI and potentially Google Gemini.
  • Apple Intelligence will be integrated into multiple devices, excluding the HomePod due to hardware limitations, and may include a new robotic device, making it comparable to iCloud in its broad application and frequent updates.

Source: https://www.techradar.com/computing/is-apple-intelligence-the-new-icloud-ai-platform-tipped-to-get-new-subscription-tier

🤖 Meta’s new dataset could enable robots to learn manual skills from human experts 

  • Meta has introduced a new benchmark dataset named HOT3D to advance AI research in 3D hand-object interactions, containing over one million frames from various perspectives.
  • This dataset aims to enhance the understanding of human hand manipulation of objects, addressing a significant challenge in computer vision research according to Meta.
  • HOT3D includes over 800 minutes of egocentric video recordings, multiple perspectives, detailed 3D pose annotations, and 3D object models, which could help robots and XR devices learn manual skills from human experts.

Source: https://the-decoder.com/metas-new-hot3d-dataset-could-enable-robots-to-learn-manual-skills-from-human-experts/

AI Innovations in June 2024

  • AI: The Ultimate Sherlocking?
    by /u/mintone (Artificial Intelligence) on July 26, 2024 at 12:16 pm

    submitted by /u/mintone [link] [comments]

  • Speech-to-Text Solution for Multilingual Sentences / Mixed-language speech
    by /u/simbaninja33 (Artificial Intelligence Gateway) on July 26, 2024 at 11:54 am

    I am looking for a speech-to-text solution, either paid or open-source, that can accurately transcribe speech containing a mix of two languages within the same sentence. I have explored options like Microsoft Azure, Google Cloud, and OpenAI, but haven't found a satisfactory solution yet. For example, I need the solution to handle sentences like: "I have tried the restaurant yesterday, it is muy muy bueno, they serve some of the pizza, que haria mi abuela super celoza de la receta." "I went to the store y compré un poco de pan because we were running low." I have already tried Microsoft Azure, which can handle multiple languages, but only when they are not mixed within the same sentence (as mentioned in their documentation). Google Cloud's speech-to-text fails to accurately transcribe mixed-language speech, and OpenAI doesn't seem to offer this functionality. I am open to both continuous real-time speech recognition and file-based recognition. For real-time applications, I am also willing to consider workarounds, such as implementing a "button" that can be clicked to quickly switch between the main language and the second language. If anyone has experience with a solution that can handle this type of mixed-language speech recognition, I would greatly appreciate any suggestions or recommendations. Thank you in advance for your help! submitted by /u/simbaninja33 [link] [comments]

  • Any open source AI model with web search abilities?
    by /u/david8840 (Artificial Intelligence Gateway) on July 26, 2024 at 11:45 am

    Is there any open source AI model with web search abilities? I want to be able to ask it questions which require real time internet searching, for example "What is the weather like now in NY?" submitted by /u/david8840 [link] [comments]

  • Which companies are leading the way in AI detection? (for audio/video deepfakes, etc.?)
    by /u/ProfessionalHat3555 (Artificial Intelligence Gateway) on July 26, 2024 at 11:21 am

    So I was listening to the most recent Bill Simmons pod w/ Derek Thompson where they discuss conspiracy theories and AI shit-detection (40:00-48:00 if you're curious)... 1ST Q: what companies are you aware of that are already working on AI detection? 2ND Q: where do you think the AI detection slice of the market is going? Will there be consumer-grade products that we can use to run, say, a political video through a detection software & get a % of realness rating on it? Will these tools ONLY be available to big conglomerates who become the purveyors of truth? 3RD Q: If we're UNABLE to do this at-scale yet, what would need to happen tech-wise for AI detection to become more accessible to more people? (disclaimer: I'm not a dev) submitted by /u/ProfessionalHat3555 [link] [comments]

  • AI can't take people's jobs if there's no people.
    by /u/baalzimon (Artificial Intelligence Gateway) on July 26, 2024 at 10:53 am

    Looks more and more likely that human populations will decline in the future. Maybe the workforce will just be AI robots rather than young people. PEW: The Experiences of U.S. Adults Who Don’t Have Children 57% of adults under 50 who say they’re unlikely to ever have kids say a major reason is they just don’t want to; 31% of those ages 50 and older without kids cite this as a reason they never had them https://www.pewresearch.org/social-trends/2024/07/25/the-experiences-of-u-s-adults-who-dont-have-children/ submitted by /u/baalzimon [link] [comments]

  • UK School Under Fire for Unlawful Facial-Recognition Use
    by /u/Think_Cat1101 (Artificial Intelligence Gateway) on July 26, 2024 at 10:43 am

    https://www.msn.com/en-us/news/technology/uk-school-under-fire-for-unlawful-facial-recognition-use/ar-BB1qEmeX?cvid=6dfe65854c6e4c2ad473b0e649e795b2&ei=10 submitted by /u/Think_Cat1101 [link] [comments]

  • OpenAI reveals 'SearchGPT'
    by /u/Mindful-AI (Artificial Intelligence Gateway) on July 26, 2024 at 10:41 am

    submitted by /u/Mindful-AI [link] [comments]

  • Amazon’s AI Chip Revolution: How They’re Ditching Nvidia’s High Prices and Speeding Ahead
    by /u/alyis4u (Artificial Intelligence Gateway) on July 26, 2024 at 9:23 am

    Six engineers tested a brand-new, secret server design on a Friday afternoon in Amazon.com’s chip lab in Austin, Texas. Amazon executive Rami Sinno said on Friday during a visit to the lab that the server was full of Amazon’s AI chips, which compete with Nvidia’s chips and are the market leader.https://theaiwired.com/amazons-ai-chip-revolution-how-theyre-ditching-nvidias-high-prices-and-speeding-ahead/ submitted by /u/alyis4u [link] [comments]

  • OpenAI's SearchGPT Is Coming For Google Search; Here Are The Features That Will Reportedly Make It Better
    by /u/vinaylovestotravel (Artificial Intelligence Gateway) on July 26, 2024 at 9:00 am

    Dubbed "SearchGPT," the tool will offer "fast and timely answers with clear and relevant sources" by referencing content from websites and news publishers, including OpenAI content partners such as News Corp (The Post's parent company) and The Atlantic. Read more: https://www.ibtimes.co.uk/openais-searchgpt-coming-google-search-here-are-features-that-will-reportedly-make-it-better-1725770 submitted by /u/vinaylovestotravel [link] [comments]

  • Deleting chats from Blackbox AI?
    by /u/Intelligent-Fig-7791 (Artificial Intelligence Gateway) on July 26, 2024 at 7:40 am

    How on earth do you delete chats from blackbox.ai ? it seems like all chats are public by default submitted by /u/Intelligent-Fig-7791 [link] [comments]

Ace the AWS Certified Data Engineer Exam

Ace the AWS Certified Data Engineer Exam (DEA-C01): Mastering AWS Services for Data Ingestion, Transformation, and Pipeline Orchestration

Ace the AWS Certified Data Engineer Exam DEA-C01: Mastering AWS Services for Data Ingestion, Transformation, and Pipeline Orchestration.

Unlock the full potential of AWS and elevate your data engineering skills with “Ace the AWS Certified Data Engineer Exam.” This comprehensive guide is tailored for professionals seeking to master the AWS Certified Data Engineer – Associate certification. Authored by Etienne Noumen, a seasoned Professional Engineer with over 20 years of software engineering experience and 5+ years specializing in AWS data engineering, this book provides an in-depth and practical approach to conquering the certification exam.

Inside this book, you will find:

Detailed Exam Coverage: Understand the core AWS services related to data engineering, including data ingestion, transformation, and pipeline orchestration.

Practice Quizzes: Challenge yourself with practice quizzes designed to simulate the actual exam, complete with detailed explanations for each answer.

Real-World Scenarios: Learn how to apply AWS services to real-world data engineering problems, ensuring you can translate theoretical knowledge into practical skills.

Hands-On Labs: Gain hands-on experience with step-by-step labs that guide you through using AWS services like AWS Glue, Amazon Redshift, Amazon S3, and more.

Expert Insights: Benefit from the expertise of Etienne Noumen, who shares valuable tips, best practices, and insights from his extensive career in data engineering.

This book goes beyond rote memorization, encouraging you to develop a deep understanding of AWS data engineering concepts and their practical applications. Whether you are an experienced data engineer or new to the field, “Ace the AWS Certified Data Engineer Exam” will equip you with the knowledge and skills needed to excel.

Prepare to advance your career, validate your expertise, and become a certified AWS Data Engineer. Embrace the journey of learning, practice consistently, and master the tools and techniques that will set you apart in the rapidly evolving world of cloud data solutions.

Get your copy today and start your journey towards AWS certification success!

Ace the AWS Certified Data Engineer Exam (DEA-C01): Mastering AWS Services for Data Ingestion, Transformation, and Pipeline Orchestration
Ace the AWS Certified Data Engineer Exam (DEA-C01): Mastering AWS Services for Data Ingestion, Transformation, and Pipeline Orchestration

Get the Ace AWS DEA-C01 Exam  eBook at Djamgatech: https://djamgatech.com/product/ace-the-aws-certified-data-engineer-exam-ebook

Get the Ace AWS DEA-C01 Exam  eBook at Google: https://play.google.com/store/books/details?id=lzgPEQAAQBAJ

Get the Ace AWS DEA-C01 Exam  eBook at Apple: https://books.apple.com/ca/book/ace-the-aws-certified-data-engineer-associate/id6504572187

Get the Ace AWS DEA-C01 Exam  eBook at Etsy: https://www.etsy.com/ca/listing/1749511877/ace-the-aws-certified-data-engineer-exam

Get the Ace AWS DEA-C01 Exam  eBook at Shopify: https://djamgatech.myshopify.com/products/ace-the-aws-certified-data-engineer-exam

The FREE Android App for AWS Certified Data Engineer Associate Exam Preparation is out and available at: https://play.google.com/store/apps/details?id=app.web.awsdataengineer.twa

Sample Quiz:

Practice Quiz 1:

A finance company is storing paid invoices in an Amazon S3 bucket. After the invoices are uploaded, an AWS Lambda function uses Amazon Textract to process the PDF data and persist the data to Amazon DynamoDB. Currently, the Lambda execution role has the following S3 permission:

{

    “Version”: “2012-10-17”,

    “Statement”: [


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

        {

            “Sid”: “ExampleStmt”,

            “Action”: [“s3:*”],

            “Effect”: “Allow”,

            “Resource”: [“*”]

        }

    ]

}

The company wants to correct the role permissions specific to Amazon S3 according to security best practices.

Which solution will meet these requirements?

  • Append “s3:GetObject” to the Action. Append the bucket name to the Resource.
  • Modify the Action to be “s3:GetObjectAttributes.” Modify the Resource to be only the bucket name.
  • Append “s3:GetObject” to the Action. Modify the Resource to be only the bucket ARN.
  • Modify the Action to be: “s3:GetObject.” Modify the Resource to be only the bucket ARN.

Practice Quiz 1 – Correct Answer: D. 

According to the principle of least privilege, permissions should apply only to what is necessary. The Lambda function needs only the permissions to get the object. Therefore, this solution has the most appropriate modifications.

Learn more about least-privilege permissions.

Practice Quiz 2:

A data engineer is designing an application that will transform data in containers managed by Amazon Elastic Kubernetes Service (Amazon EKS). The containers run on Amazon EC2 nodes. Each containerized application will transform independent datasets and then store the data in a data lake. Data does not need to be shared to other containers. The data engineer must decide where to store data before transformation is complete.

Which solution will meet these requirements with the LOWEST latency?

  • Containers should use an ephemeral volume provided by the node’s RAM.
  • Containers should establish a connection to Amazon DynamoDB Accelerator (DAX) within the application code.
  • Containers should use a PersistentVolume object provided by an NFS storage.
  • Containers should establish a connection to Amazon MemoryDB for Redis within the application code.

Practice Quiz 2 – Correct Answer: A. 

Amazon EKS is a container orchestrator that provides Kubernetes as a managed service. Containers run in pods. Pods run on nodes. Nodes can be EC2 instances, or nodes can use AWS Fargate. Ephemeral volumes exist with the pod’s lifecycle. Ephemeral volumes can access drives or memory that is local to the node. The data does not need to be shared, and the node provides storage. Therefore, this solution will have lower latency than storage that is external to the node.

Learn more about Amazon EKS storage.

Learn more about persistent storage for Kubernetes.

Learn more about EC2 instance root device volume.

Learn more about Amazon EKS nodes.

Ace the AWS Certified Data Engineer Exam
Ace the AWS Certified Data Engineer Exam

Resources and Tips:

Ace the AWS DEA-C01 GPT:

Ace the AWS Certified Data Engineer Exam GPT
Ace the AWS Certified Data Engineer Exam GPT

Courses

New! Sessions on Twitch by AWS DevRel teams focused on DEA Exam

See : https://pages.awscloud.com/GLOBAL-other-T2-Traincert-AWS-Power-Hour-Data-Engineer-Associate-Season1-2024-reg.html

Free beginner level courses from AWS Skill builder.

Fundamentals of Data Analytics on AWS – the current Skillbuilder course linked from the certification site is ending in Feb and is being split into 2 new courses :

https://explore.skillbuilder.aws/learn/course/internal/view/elearning/18437/fundamentals-of-analytics-on-aws-part-1

https://explore.skillbuilder.aws/learn/course/internal/view/elearning/18440/fundamentals-of-analytics-on-aws-part-2

Ace the AWS Certified Data Engineer Exam Book Preview

Ace the AWS Certified Data Engineer Exam: Mastering AWS Services for Data Ingestion, Transformation, and Pipeline Orchestration
Ace the AWS Certified Data Engineer Exam: Mastering AWS Services for Data Ingestion, Transformation, and Pipeline Orchestration

Wonderland Bedtime Adventures: Diverse Tales for Dreamy Nights

Wonderland Bedtime Adventures: Diverse Tales for Dreamy Nights

Wonderland Bedtime Adventures: Diverse Tales for Dreamy Nights.

Welcome to Wonderland Bedtime Stories, where diversity shines through every tale. Our multilingual stories celebrate the rich heritage of children from all corners of the globe, ensuring every child sees themselves in the magic of storytelling.

Embark on a magical journey with “Wonderland Bedtime Adventures: Diverse Tales for Dreamy Nights” by Etienne Noumen. This enchanting collection of bedtime stories is designed to celebrate the beauty of diversity and inclusivity. With a rich blend of traditional and modern tales, this book introduces young readers to heroes and heroines from all cultures, shapes, sizes, and backgrounds.

Etienne Noumen, a seasoned software engineer and father of four, leverages cutting-edge AI to bring these stories to life. Each tale is crafted to spark imagination, foster a love for reading, and create a special bond during bedtime. From exciting adventures to heartwarming narratives, these stories are perfect for snuggling up with your little ones and drifting off to dreamland.

“Wonderland Bedtime Adventures” is more than just a collection of stories; it’s a celebration of education, sports, and the power of inclusion. Ideal for parents seeking diverse and inspiring tales, this book ensures every child sees themselves reflected in the magic of storytelling.

Key Features:

A collection of diverse and inclusive bedtime stories

Perfect for children of all ages

Promotes cultural awareness and inclusivity

Uses AI technology to blend traditional and new tales

Written by Etienne Noumen, a passionate storyteller and father

Join us in “Wonderland Bedtime Adventures” and discover a world where every child’s dreams are celebrated and cherished. Perfect for bedtime, naptime, or anytime a story is needed!

Foreword: A World of Stories, A World of You

Hello, and welcome to Wonderland Bedtime Stories! My name is Etienne Noumen, and I’m a software engineer with a passion for storytelling and a heart full of bedtime memories.

As a father of four, reading to my children has been one of life’s greatest joys. It’s a chance to spark their imaginations, introduce them to new cultures, and create a special bond before drifting off to dreamland.

But something always bothered me. Many of the bedtime stories we read lacked the beautiful diversity of the world we live in. Where were the heroes and heroines who looked like my children, with their unique backgrounds and dreams?

This yearning for inclusivity is the seed from which Wonderland Bedtime Stories blossomed. As a software engineer with over 20 years of experience, I saw the potential of technology to bridge this gap. Here, we use cutting-edge AI to recreate traditional tales and weave brand new stories that celebrate the richness of different cultures and ethnicities.

Wonderland Bedtime Stories is more than just bedtime tales; it’s a world where every child can see themselves reflected in the magic of storytelling. It’s a journey to faraway lands, where heroes come in all shapes, sizes, and skin tones. It’s a celebration of education, sports, and the power of inclusion – values I hold dear as a father and a human being.

So, snuggle up with your little ones, open a book, and embark on a wondrous adventure together. We believe that within these pages, every child will find their own unique reflection, a spark of inspiration, and a world of dreams waiting to be explored.

Happy Reading,

Etienne Noumen, P.Eng

Founder, Wonderland Bedtime Stories

The Whispering Baobab

What is the story about:

In a village by the savannah, Nia, a curious girl, was drawn to an ancient baobab, known as the Whispering Baobab. It whispered wisdom from ages past. One evening, the tree shared its stories with Nia, telling of wildlife, seasons, and village life beneath its branches. It imparted lessons of respect and harmony with nature. Entrusted with these tales, Nia became the village storyteller, spreading the baobab’s wisdom, linking past and future with each story.


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

Full Story:

The Whispering Baobab:

Wonderland Bedtime Adventures: The Whispering Baobab
Wonderland Bedtime Adventures: The Whispering Baobab

Welcome, little dreamers, to another enchanting episode of “Wonderland Bedtime Adventures.” Snuggle up in your cozy blankets as we journey to a magical village on the edge of a sprawling savannah. Our tale tonight is about a young girl named Nia and a very special tree. Are you ready? Let’s begin. Once upon a time, in a world not too far from our own, there was a small village where the sun kissed the earth goodnight with rays of gold. In this village lived a young girl named Nia, known for her sparkling eyes and a heart full of wonder. Nia’s favorite place to be was outside, exploring the vast lands that stretched beyond her home. Her most beloved spot was under an ancient baobab tree, grand and wise. Its massive trunk was wider than any house in the village, and it was called the Whispering Baobab. The villagers would often say, with a twinkle in their eye, that this tree whispered the wisdom of ages to those who would listen. And listen Nia did! She visited the baobab every day, hoping to catch its mystical murmurs. One warm evening, as the world turned honeyed hues of sunset, Nia sat under the great tree. The savannah was alive with the wild calls of animals, and the baobab’s leaves played a gentle song in the breeze. That’s when it happened. A faint whisper tickled Nia’s ear. “Welcome, Nia,” said the baobab, its voice deep and slow, like the ancient earth itself. Nia’s eyes widened with joy as she answered, “I knew you were special, Whispering Baobab. Please share your stories with me.” Oh, the tales the baobab told! Stories of majestic wildlife, of changing seasons, and of children, just like yourselves, playing under its branches. It spoke of tough times and rain-filled days, of laughter, and the dance of life on the savannah. The baobab also taught Nia something very important. It told her to respect nature, to preserve the environment, and to live peacefully with all creatures. As the stars blinked sleepily above, the baobab whispered a secret to Nia, “You are now the keeper of my stories. Share them, teach them, for our lives are as interconnected as the roots of a baobab tree.” With the first light of dawn, Nia returned to her village, ready to share the baobab’s wisdom. She became the Storyteller of the Savannah, a title she wore with pride, spreading the whispers and the love for nature to everyone. And so, dear listeners, as you drift into your dreams tonight, imagine the wide, starry skies of the savannah and the ancient Whispering Baobab sharing its secrets with you. Remember, like the roots of the baobab, we are all connected in the great circle of life. Goodnight, and may your dreams be filled with wonderland adventures.

Conclusion: Sweet Dreams, Little Wanderers!

The world of Wonderland Bedtime Stories is now yours to explore! Fifty adventures await, each one filled with exciting journeys, brave heroes, and magical lands.

Remember, within these pages, you might meet a princess who looks just like you, a courageous explorer with your same adventurous spirit, or a wise storyteller who shares your love for learning. Every story holds a piece of the amazing world we live in, and every character reflects the beauty of our differences.

So, as you drift off to sleep, hold onto the magic of the stories you’ve read. Dream of soaring through the clouds, conquering mountains, or discovering hidden treasures. Let these stories inspire you to be kind, be brave, and believe in yourself.

Sweet dreams, little wanderers! May your dreams be filled with the wonder and excitement you’ve discovered in Wonderland Bedtime Stories.

Wonderland Bedtime Stories: Sweet Dreams, little wanderers
Wonderland Bedtime Stories: Sweet Dreams, little wanderers

Tips for Using AI for Your Rental Property

A suited man holding his hand over an iPad which has a hovering digital overlay displaying the words "Property Management."

Artificial Intelligence (AI) is transforming the rental property industry, making management tasks easier and more efficient. Here are some essential tips for using AI for your rental property to streamline operations and enhance tenant satisfaction.

Automate Routine Tasks

One of the biggest advantages of incorporating AI into your rental property management is its ability to automate routine tasks. This includes tasks such as the following:

  • Rent collection
  • Scheduling maintenance
  • Responding to common tenant inquiries
  • Sending out documents and news

Automation allows you to focus on higher-priority tasks, such as building tenant relationships and expanding your property portfolio.

Use AI for Tenant Screening

AI can significantly improve the tenant screening process by swiftly analyzing vast amounts of data to identify the best candidates. Advanced algorithms can evaluate prospective tenants based on credit scores, rental history, and employment status, providing a comprehensive assessment within seconds. For instance, AI can help you spot a phony pay stub by detecting inconsistencies or patterns indicative of fraud.

Predict Maintenance Needs

Maintenance is a critical aspect of property management that can affect tenant satisfaction and property value. AI can predict maintenance needs by analyzing data from various sensors and historical maintenance records. These predictive algorithms can forecast when equipment is likely to fail or when a specific area of the property requires attention, allowing for preemptive actions. This reduces unexpected breakdowns, lowers maintenance costs, and ensures that your property remains in excellent condition.

Optimize Property Marketing

AI technologies can also revolutionize how you market your rental properties. AI-driven tools can analyze market trends and demographics to identify the best times and platforms to list your vacancies. They can optimize your listings with dynamic pricing strategies that maximize your rental income based on current market conditions.

Additionally, AI can help create targeted advertising campaigns that reach the most promising potential tenants. By making data-driven marketing decisions, you increase the chances of reducing vacancy rates and shortening the time your properties remain unoccupied.

In short, integrating AI into your rental property management strategy offers numerous benefits. These tips for using AI for your rental property will help streamline processes, enhance tenant satisfaction, and optimize your overall operations. Embrace these technological advancements to stay competitive in the ever-evolving real estate market.

AI Innovations in June 2024

AI Innovations in June 2024

AI Innovations in June 2024.

Welcome to our blog series “AI Innovations in June 2024”! This is an evolving article that will be updated daily throughout the month of June 2024 to bring you the latest news and developments in the world of artificial intelligence. As we move further into the 21st century, the pace of AI innovation continues to accelerate at a breathtaking rate. Each day, researchers, engineers, and visionaries are pushing the boundaries of what’s possible, unlocking new capabilities and applications that are transforming industries, enhancing our lives, and shaping the future. In this blog, we’ll dive deep into the most exciting AI breakthroughs, advancements, and milestones happening in June 2024. From groundbreaking AI-powered technologies and cutting-edge research to the societal and ethical implications of these innovations, we’ll provide you with a comprehensive and insightful look at the rapidly evolving world of artificial intelligence. Whether you’re an AI enthusiast, a tech-savvy professional, or simply someone curious about the future, this blog will keep you informed, inspired, and engaged. So, join us on this journey of discovery as we explore the frontiers of AI and uncover the innovations that are shaping our world. Stay tuned for daily updates, and get ready to be amazed by the incredible advancements that are happening in the world of AI!

LISTEN DAILY AT OUR PODCAST HERE

Experience the transformative capabilities of AI with “Read Aloud For Me – AI Dashboard – AI Tools Catalog – AI Tools Recommender” – your ultimate AI Dashboard and Hub. Seamlessly access a comprehensive suite of top-tier AI tools within a single app, meticulously crafted to enhance your efficiency and streamline your digital interactions. Now available on the web at readaloudforme.com and across popular app platforms including Apple, Google, and Microsoft, “Read Aloud For Me – AI Dashboard” places the future of AI at your fingertips, blending convenience with cutting-edge innovation. Whether for professional endeavors, educational pursuits, or personal enrichment, our app serves as your portal to the forefront of AI technologies. Embrace the future today by downloading our app and revolutionize your engagement with AI tools.

AI Dashboard - Wonderland Bedtime Stories - Safe Ai for All
AI Dashboard – Wonderland Bedtime Stories – Safe Ai for All

📊Top AI Statistics 2024 Roundup.

📊Top AI Statistics 2024 Roundup.
📊Top AI Statistics 2024 Roundup.

  1. The global AI market is valued at over $196 billion.
  2. AI industry value is projected to increase by over 13x over the next 7 years.
  3. The US AI market is forecast to reach $299.64 billion by 2026.
  4. The AI market is expanding at a CAGR of 38.1% between 2022 to 2030.
  5. By 2025, as many as 97 million people will work in the AI space.
  6. AI market size is expected to grow by at least 120% year-over-year.
  7. 83% of companies claim that AI is a top priority in their business plans.
  8. Netflix makes $1 billion annually from automated personalized recommendations.
  9. 48% of businesses use some form of AI to utilize big data effectively.
  10. 38% of medical providers use computers as part of their diagnosis.

Credit: Josh Howarth

AI Weekly Summary June 22-30 2024

🤝 Apple and Meta are discussing a generative AI partnership
🔧 ByteDance and Broadcom collaborate on AI chip development
🕵️‍♂️ Researchers developed a new method to detect hallucinations
🎥 Synthesia 2.0: World’s 1st AI video communication platform
🛒 OpenAI is on an acquiring spree, buying Rocket and Multi
🎵 Record labels sue AI music startups over copyright infringement
💼 Anthropic rolls out Claude’s cutting-edge collaborative features
🤖 Google experiments with celebrity-inspired AI Chatbots
🛑 OpenAI postpones the launch of ChatGPT voice mode
🏁 Amazon steps into the chatbot race with Metis
🎨 Figma’s new AI features stir competition with Adobe
🥇 Alibaba’s Qwen-72B tops Hugging Face’s Open LLM Leaderboard
🚀 Google releases Gemma 2, lightweight but powerful open LLMs
🔍 OpenAI’s CriticGPT finds GPT-4’s mistakes with GPT-4
🌐 Google partners with Moody’s, Thomson Reuters & more for AI data

Apple and Meta are discussing gen AI partnership

Apple is reportedly in talks with its longtime rival Meta to integrate the latter’s Llama 3 AI model into Apple Intelligence. This move comes as Apple prepares to roll out its AI features across iPhones, iPads, and Macs later this year.

The potential partnership follows Apple’s existing deal with OpenAI, suggesting a collaboration strategy rather than solo development in the AI race. In Apple’s arrangement with OpenAI, there’s no direct payment. Instead, OpenAI can offer premium subscriptions through Apple Intelligence, with Apple taking a percentage. It’s unclear if Meta would agree to a similar business model, given that Llama 3 is open-source and free to access.

Sourcehttps://www.wsj.com/tech/ai/apple-meta-have-discussed-an-ai-partnership-cc57437e

ByteDance and Broadcom collaborate on AI chip development

ByteDance is collaborating with U.S. chip designer Broadcom to develop an advanced AI processor. This partnership aims to secure a stable supply of high-end chips amid ongoing U.S.-China tensions. The project centers on creating a 5-nanometre, customized Application-Specific Integrated Chip (ASIC) that complies with U.S. export restrictions.

This chip’s manufacturing is set to be outsourced to Taiwan Semiconductor Manufacturing Company (TSMC), though production is not expected to begin this year. While the design work is currently underway, the critical “tape out” phase has yet to commence.

Sourcehttps://www.reuters.com/technology/artificial-intelligence/chinas-bytedance-working-with-broadcom-develop-advanced-ai-chip-sources-say-2024-06-24

Researchers developed a new method to detect hallucinations

ChatGPT and Gemini can produce impressive results but often “hallucinate” false or unsubstantiated information. This research focuses on a subset of hallucinations called “confabulations,” where LLMs generate answers that are both wrong and arbitrary. Researchers have developed new methods to detect confabulations using entropy-based uncertainty estimators. They introduce the concept of “semantic entropy” to measure the uncertainty of LLM generations at the meaning level.

High semantic entropy corresponds to high uncertainty and indicates a higher likelihood of confabulation. The method computes uncertainty at the level of meaning rather than specific word sequences, addressing the fact that one idea can be expressed in many ways. The method provides scalable oversight by detecting confabulations that people might otherwise find plausible.

Sourcehttps://www.nature.com/articles/s41586-024-07421-0

Synthesia 2.0: World’s 1st AI video communication platform

Synthesia is launching Synthesia 2.0 – the world’s first AI video communications platform for businesses. It reinvents the entire video production process, allowing companies to create and share AI-generated videos at scale easily.

The key new features and capabilities of Synthesia 2.0 include:

  • 2 Personal AI Avatars: Expressive Avatars shot in a studio and Custom Avatars created using your webcam.
  • AI Video Assistant: Converts text, documents, or websites into high-quality videos, with options to customize the branding, tone, and length.
  • Intuitive Video Editing: Editing simplified with “Triggers” that let you control animations and edits from the script.
  • Translation and Dynamic Video Player: Videos can now be translated into over 120 languages. Synthesia is also building a new video player with interactive features.
  • AI Safety Focus: Synthesia is pursuing ISO/IEC 42001 certification, the first standard for responsible AI management, to ensure its AI technologies are ethical.

Sourcehttps://www.synthesia.io/post/introducing-synthesia-video-communications-platform?

OpenAI is on an acquiring spree, buying Rockset and Multi

Last week, OpenAI acquired Rockset, a startup that develops tools for real-time data search and analytics. OpenAI said it would integrate Rockset’s technology to power its infrastructure and offerings across products.


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

This week, OpenAI acquired Multi, a startup focused on building remote collaboration tools and software. Technically, the deal is an acqui-hire as the entire Multi team, including its co-founders, will join OpenAI to work on the company’s ChatGPT desktop application.

Sourcehttps://techcrunch.com/2024/06/24/openai-buys-a-remote-collaboration-platform

Record labels sue AI music startups over copyright infringement

The world’s major record labels, including Universal Music Group, Sony Music, and Warner Music, have filed twin lawsuits against the AI music generation startups Suno and Udio. The lawsuits accuse the companies of unlawfully training their AI models on massive amounts of copyrighted music, which, according to the complaints, allows the startups to generate similar-sounding music without permission.

The record labels allege Suno and Udio have effectively copied artists’ styles and specific musical characteristics. The labels claim the AI-generated music is so close to the original that it is eerily similar when transcribed into sheet music. The lawsuits also accuse the startups of making it easy for people to distribute AI-created samples that mimic copyrighted recordings on platforms like Spotify.

Sourcehttps://venturebeat.com/ai/record-labels-sue-ai-music-generator-startups-suno-udio-for-copyright-infringement/

Anthropic rolls out Claude’s cutting-edge collaborative features

Anthropic has introduced new collaboration features for Claude. These features include:

  • Projects: Projects in Claude allow integration of internal resources like style guides or codebases, enhancing Claude’s ability to deliver tailored assistance across various tasks. Users can set custom instructions for each Project to modify Claude’s tone or perspective for a specific role or industry.
  • Artifacts: It allows users to generate and edit various content types like code, documents, and graphics within a dedicated window. This benefits developers by offering larger code windows and live previews for easier front-end reviews.
  • Sharing Features: Claude Team users can share snapshots of their best conversations with Claude in their team’s shared project activity feed.

Additionally, any data or chats shared within Projects will not be used to train Anthropic’s generative models without a user’s explicit consent.

Sourcehttps://www.anthropic.com/news/projects

Google experiments with celebrity-inspired AI Chatbots

These chatbots will be powered by Google’s Gemini family of LLMs. The company aims to strike partnerships with influencers and celebrities and is also working on a feature that allows people to create their own chatbots by describing their personalities and appearances.

The project is led by Ryan Germick, a longtime executive at Google and a team of ten. These chatbots could be an experiment and may only appear on Google Labs rather than being widely available.

Sourcehttps://www.msn.com/en-us/news/other/google-wants-to-build-ai-chatbots-based-on-celebs-influencers-for-some-reason/ar-BB1oS1or

OpenAI postpones the launch of ChatGPT voice mode

Originally planned for late June, the Voice Mode aims to provide a more naturalistic and conversational experience with the AI chatbot, complete with emotional inflection and the ability to handle interruptions.

However, it will now be available only to a small group of users in late July or early August. OpenAI is working on improving content detection and user experience before wider rollout. GPT-4o’s real-time voice and vision capabilities are also expected to roll out to ChatGPT Plus users soon.

Sourcehttps://techcrunch.com/2024/06/25/openai-delays-chatgpts-new-voice-mode

Amazon steps into the chatbot race

Amazon is reportedly working on a new consumer-focused chatbot codenamed “Metis.” It is planned to be released somewhere around September. Here’s what we know about it:

  • The chatbot is powered by a new model, Olympus, and can be accessed via a web browser.
  • It uses a retrieval-augmented generation (RAG) technique to provide up-to-date information and automate tasks.
  • The model conversationally provides text and image-based outputs, suggesting follow-ups to queries. It also shares links to sources and supports image generation.
  • It uses an infrastructure similar to Amazon’s upcoming voice assistant, Remarkable Alexa.

Sourcehttps://www.businessinsider.com/amazon-chatgpt-rival-codenamed-metis-2024-6

Figma’s new AI features stir competition with Adobe

Figma announced a range of new features at the 2024 Config conference. Significant ones include a UI redesign, generative AI tools, new icons and toolbar, AI-enhanced asset search, and auto-generated texts in designs.

For instance, by typing a simple prompt into the textbox, users can create an entire app design mock-up for a restaurant. Figma will connect the design pages and even write suggested content!

Figma has also added a few designer-specific features to allow users to tweak designs in real-time. It features a developer mode with a “ready-for-dev” task list. The upgrade also boasts Figma slides, a Google slides-like tool for building and sharing presentations.

Sourcehttps://www.figma.com/whats-new/

Alibaba’s Qwen-72B tops the Hugging Face leaderboard

Hugging Face’s latest open large language model leaderboard ranks and evaluates open LLMs based on benchmarks like MMLU-pro and tests them on high-school and college-level problems.

The platform used 300 NVIDIA H100 GPUs to re-evaluate major open LLMs to obtain updated rankings. Chinese company Alibaba’s Qwen-72B dominated the leaderboard, becoming a top performer overall.

Not just that, the leaderboard was mainly dominated by Chinese companies, highlighting their headway into the open LLM space.

Sourcehttps://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard

Google’s Gemma 2, a set of lightweight, powerful open LLMs

Google has released Gemma 2 set of models that punch above their weight classes. Available in 9B and 27B parameter sizes, these models are

  • Higher performing and more efficient at inference than the first-generation
  • Have significant safety advancements built in
  • Optimized to run at incredible speed across a range of hardware and easily integrate with other AI tools
  • Trained on 13 trillion tokens for 27B, 8 trillion for 9B, and 2 trillion for 2.6B model (en route)

27B performs better than Llama3-70B and Nemotron-340B on Lmsys Arena, making it best in its size and stronger than some larger models. While 9B outperforms the likes of Mistral-large and Qwen1.5-110B.

The 27B Gemma 2 model is designed to run inference efficiently at full precision on a single Google Cloud TPU host, NVIDIA A100 80GB Tensor Core GPU, or NVIDIA H100 Tensor Core GPU. Moreover, this is an open weights model line, currently only available to researchers and developers.

Sourcehttps://blog.google/technology/developers/google-gemma-2

OpenAI’s CriticGPT finds GPT-4’s mistakes with GPT-4

OpenAI trained a model based on GPT-4, called CriticGPT, to catch errors in ChatGPT’s code output. It found that when users get help from CriticGPT to review ChatGPT code, they outperform those without help 60% of the time.

Ace the Microsoft Azure Fundamentals AZ-900 Certification Exam: Pass the Azure Fundamentals Exam with Ease

OpenAI aligns GPT-4 models to be more helpful and interactive through Reinforcement Learning from Human Feedback (RLHF). A key part of RLHF is collecting comparisons in which people, called AI trainers, rate different ChatGPT responses against each other.

OpenAI is beginning to integrate CriticGPT-like models into its RLHF labeling pipeline, providing trainers with explicit AI assistance.

Sourcehttps://openai.com/index/finding-gpt4s-mistakes-with-gpt-4

Google’s partnerships to help AI with real-world facts

Google is partnering with reputable third-party services, such as Moody’s, MSCI, Thomson Reuters, and Zoominfo, to ground its AI with real-world data. These four will be available within Vertex AI starting next quarter. They will offer developers qualified data to backstop their model outputs and ensure responses are factually accurate.

Google is also announcing high-fidelity grounding. Available through an experimental preview, it’s designed to help AI systems work better with a given set of specific information.

Sourcehttps://venturebeat.com/ai/google-grounding-ai-with-moodys-msci-thomson-reuters-zoominfo

A  Daily chronicle of AI Innovations June 28th 2024:

🚀 Google releases Gemma 2, a set of lightweight but powerful open LLMs
🔍 OpenAI’s CriticGPT finds GPT-4’s mistakes with GPT-4
🌐 Google partners with Moody’s, Thomson Reuters & more for AI data

🔎 OpenAI’s critics train smarter AI systems

🤝TIME and OpenAI announced a multi-year content deal

🌍Google is using AI to add 110 new languages to Google Translate

🎼YouTube is in talks with major record labels for an AI music deal

🤖Meta to start testing user-created AI chatbots on Instagram

📞Character.AI now allows users to talk with AI avatars over calls

🔎 OpenAI’s critics train smarter AI systems

OpenAI just published new research exploring how to improve model training using AI ‘critics’ that help human trainers spot errors in AI outputs more effectively.

  • OpenAI developed a model called CriticGPT to assist human trainers in evaluating and catching errors in code written by AI systems like ChatGPT.
  • In tests, CriticGPT caught 85% of inserted bugs in AI-generated code, compared to just 50% for human reviewers alone.
  • Human trainers also preferred CriticGPT’s code critics over their own 63% of the time on naturally occurring bugs.
  • OpenAI said that CriticGPT-type models are already being integrated into the company’s training workflow, helping prepare for advancing AI systems.

As OpenAI states in the research paper, both LLMs and AI critics will continue to scale — human intelligence will not. The limitations of our own puny mortal brains look to be no match for evaluating systems that are continuing to soar, meaning AI is quickly becoming its own trainer.

Source: https://openai.com/index/finding-gpt4s-mistakes-with-gpt-4

🚀 Google releases Gemma 2, a set of lightweight but powerful open LLMs

Google has released Gemma 2 set of models that punch above their weight classes. Available in 9B and 27B parameter sizes, these models are

If you are looking for an all-in-one solution to help you prepare for the AWS Cloud Practitioner Certification Exam, look no further than this AWS Cloud Practitioner CCP CLF-C02 book

  • Higher performing and more efficient at inference than the first-generation
  • Have significant safety advancements built in
  • Optimized to run at incredible speed across a range of hardware and easily integrate with other AI tools
  • Trained on 13 trillion tokens for 27B, 8 trillion for 9B, and 2 trillion for 2.6B model (en route)

27B performs better than Llama3-70B and Nemotron-340B on Lmsys Arena, making it best in its size and stronger than some larger models. While 9B outperforms the likes of Mistral-large and Qwen1.5-110B.

The 27B Gemma 2 model is designed to run inference efficiently at full precision on a single Google Cloud TPU host, NVIDIA A100 80GB Tensor Core GPU, or NVIDIA H100 Tensor Core GPU. Moreover, this is an open weights model line, currently only available to researchers and developers.

Why does it matter?

The models sound like they are built for practical deployments. They come in practical sizes so that they can be easily deployed while being amazing in quality due to best-in-class performances.

Source: https://blog.google/technology/developers/google-gemma-2/

🔍 OpenAI’s CriticGPT finds GPT-4’s mistakes with GPT-4

OpenAI trained a model based on GPT-4, called CriticGPT, to catch errors in ChatGPT’s code output. It found that when users get help from CriticGPT to review ChatGPT code, they outperform those without help 60% of the time.

OpenAI aligns GPT-4 models to be more helpful and interactive through Reinforcement Learning from Human Feedback (RLHF). A key part of RLHF is collecting comparisons in which people, called AI trainers, rate different ChatGPT responses against each other.

OpenAI is beginning to integrate CriticGPT-like models into its RLHF labeling pipeline, providing trainers with explicit AI assistance.

Why does it matter?

With more advances in reasoning and model behavior, AI models’ mistakes can become more subtle for AI trainers to spot. CriticGPT is a step towards addressing this fundamental limitation of RLHF.

Source: https://openai.com/index/finding-gpt4s-mistakes-with-gpt-4

🌐 Google partners with Moody’s, Thomson Reuters & more for AI data

Google is partnering with reputable third-party services, such as Moody’s, MSCI, Thomson Reuters, and Zoominfo, to ground its AI with real-world data. These four will be available within Vertex AI starting next quarter. They will offer developers qualified data to backstop their model outputs and ensure responses are factually accurate.

Google is also announcing high-fidelity grounding. Available through an experimental preview, it’s designed to help AI systems work better with a given set of specific information.

Why does it matter?

Earlier, Google announced efforts to ground Vertex AI results using web data and a plan to allow companies to ground AI systems in their own internal data.

Now, it is grounding these systems in known factual data from third parties, which could significantly lessen hallucinations and make AI more trustworthy for enterprise customers.

Source: https://venturebeat.com/ai/google-grounding-ai-with-moodys-msci-thomson-reuters-zoominfo

What Else Is Happening in AI on June 27th 2024❗

🤝TIME and OpenAI announced a multi-year content deal

OpenAI will gain access to current and historic content from TIME’s extensive archives from the last 101 years to enhance its products. It will also enable TIME to gain access to OpenAI’s technology to develop new products for its audiences.

Source: https://openai.com/index/strategic-content-partnership-with-time

🌍Google is using AI to add 110 new languages to Google Translate

It is Google’s largest expansion ever, thanks to its PaLM 2 LLM. It includes languages like Cantonese, NKo, and Tamazight, representing more than 614 million speakers and opening up translations for around 8% of the world’s population.

Source: https://blog.google/products/translate/google-translate-new-languages-2024/

🎼YouTube is in talks with major record labels for an AI music deal

It is offering to pay Universal Music Group (UMG), Sony Music Entertainment, and Warner Records “lump sums of cash” in exchange for legally licensing their songs to train new AI music tools. These will likely be one-off payments, not royalty-based arrangements. Source: https://www.theverge.com/2024/6/27/24187151/youtube-ai-music-deals-licensing-record-labels-sony-umg-warner

🤖Meta to start testing user-created AI chatbots on Instagram

CEO Mark Zuckerberg announced yesterday that Meta will begin to surface AI characters made by creators through Meta AI studio on Instagram, starting in the U.S. These will primarily show up in messaging for now and will be clearly labeled as AI.

Source: https://techcrunch.com/2024/06/27/meta-starts-testing-user-created-ai-chatbots-on-instagram

📞Character.AI now allows users to talk with AI avatars over calls

Users can initiate calls with a user-generated AI character directly with a button tap. Users can also switch between calling and texting seamlessly and stop the AI from talking through a “Tap to interrupt” option. The feature currently supports only a few languages.

Source: https://www.reuters.com/technology/artificial-intelligence/ai-chatbot-startup-characterai-launches-new-calls-feature-2024-06-27

📊 Use ChatGPT to turn docs into spreadsheets

With a simple prompt, ChatGPT can analyze documents, answer questions, perform calculations, and create a downloadable spreadsheet — all in one conversation!

  1. Log in to ChatGPT and upload your document to the chat (remember to remove any sensitive data before submitting).
  2. Ask questions about it, e.g., “How much is the security deposit?”
  3. Request calculations, like the total first month’s costs, including deposits and fees.
  4. Simple prompt ChatGPT to “create a downloadable budget spreadsheet”, specifying the time frame and desired columns.

Note: Only share documents you’re comfortable with. Be cautious with sensitive information and consider redacting critical details before uploading.

Source: https://university.therundown.ai/c/daily-tutorials/from-lease-agreement-to-budget-spreadsheet-with-chatgpt-0ed8d7ba-7d5a-4e9f-903a-64822caf1579

A  Daily chronicle of AI Innovations June 27th 2024:

🎵 YouTube wants to pay record labels to use their songs for AI training

🎨 Figma unveils AI-powered design tools, challenges Adobe’s dominance

🗣️ Character.AI introduces AI avatar calls

🏁 Amazon steps into the chatbot race 

🥇 Alibaba’s Qwen-72B tops the Hugging Face leaderboard

🏆 Hugging Face updates Open LLM Leaderboard

🎙️ NBC to use AI voice for Olympics recaps!

💡 Sonia’s ‘AI Therapist’ sparks debate

🏛️ Anthropic extends Claude’s access to government

🎵 YouTube wants to pay record labels to use their songs for AI training

  • YouTube is reportedly negotiating with major music labels like Sony Music Entertainment, Universal Music Group, and Warner Records to pay for licensing their songs for AI training.
  • Despite these negotiations, the ultimate decision to allow the use of songs for AI training will likely fall to individual artists rather than the labels themselves.
  • In April 2023, more than 200 artists voiced their opposition to AI using their work, emphasizing the potential threat to creators’ rights and the music ecosystem.

Source: https://www.engadget.com/youtube-reportedly-wants-to-pay-record-labels-to-use-their-songs-for-ai-training-125052503.html

🎨 Figma unveils AI-powered design tools, challenges Adobe’s dominance

  • Figma has launched AI-powered design features to speed up workflows and aid teams in creating superior software, positioning it as a product development platform to compete with Adobe.
  • The new Figma AI features include design creation with text prompts, visual search capabilities, layer renaming, and automating repetitive tasks, and are currently available in a limited beta.
  • While beta access to Figma AI is free, later availability will require subscriptions to specific plans, and admins must enable AI features, differentiating Figma’s approach from Adobe’s controversial AI terms.

Source: https://venturebeat.com/ai/figma-unveils-ai-powered-design-tools-to-rival-adobe/

🗣️ Character.AI introduces AI avatar calls

  • Character.AI announced that users can now make calls to AI characters, supporting multiple languages including English, Spanish, Portuguese, Russian, Korean, Japanese, and Chinese.
  • The company reported that during testing, over 3 million users made 20 million calls, and noted that AI character calls can help with language practice, mock interviews, or role-playing games.
  • With a button tap, users can start calls with AI characters, seamlessly switch between calling and texting, and use a “Tap to interrupt” feature to stop the AI from talking.

Source: https://techcrunch.com/2024/06/27/character-ai-now-allows-users-to-talk-with-avatars-over-calls/

🏁 Amazon steps into the chatbot race

Amazon is reportedly working on a new consumer-focused chatbot codenamed “Metis.” It is planned to be released somewhere around September. Here’s what we know about it:

  • The chatbot is powered by a new model, Olympus, and can be accessed via a web browser.
  • It uses a retrieval-augmented generation (RAG) technique to provide up-to-date information and automate tasks.
  • The model conversationally provides text and image-based outputs, suggesting follow-ups to queries. It also shares links to sources and supports image generation.
  • It uses an infrastructure similar to Amazon’s upcoming voice assistant, Remarkable Alexa.

Why does it matter?

The AI assistant market is already too crowded, making Metis’ release seem a bit too late. Its rivals, OpenAI and Google, have already launched their models in 2023 and will give tough competition to Amazon. If Amazon wishes to win this AI race, Metis must showcase exceptional capabilities.

Source: https://www.businessinsider.com/amazon-chatgpt-rival-codenamed-metis-2024-6

 
🎨 Figma’s new AI features stir competition with Adobe

Figma announced a range of new features at the 2024 Config conference. Significant ones include a UI redesign, generative AI tools, new icons and toolbar, AI-enhanced asset search, and auto-generated texts in designs.

https://x.com/figma/status/1806016455609286847

For instance, by typing a simple prompt into the textbox, users can create an entire app design mock-up for a restaurant. Figma will connect the design pages and even write suggested content!

Figma has also added a few designer-specific features to allow users to tweak designs in real-time. It features a developer mode with a “ready-for-dev” task list. The upgrade also boasts Figma slides, a Google slides-like tool for building and sharing presentations.

Why does it matter?

Figma’s AI tools will enable newer users on the platform to test out ideas easily. At the same time, tools like “ready-for-dev” will offer quick iterations to users who are more familiar with the platform.

With Adobe abandoning its plans to acquire Figma, the company’s recent upgrade might pose a challenge to the former.

Source: https://www.figma.com/whats-new/

🥇 Alibaba’s Qwen-72B tops the Hugging Face leaderboard

Hugging Face’s latest open large language model leaderboard ranks and evaluates open LLMs based on benchmarks like MMLU-pro and tests them on high-school and college-level problems.

The platform used 300 NVIDIA H100 GPUs to re-evaluate major open LLMs to obtain updated rankings. Chinese company Alibaba’s Qwen-72B dominated the leaderboard, becoming a top performer overall.

Not just that, the leaderboard was mainly dominated by Chinese companies, highlighting their headway into the open LLM space.

Why does it matter?

The open LLM leaderboard is essential for transparent and comprehensive LLM evaluation. It also proves that a model’s size doesn’t guarantee its superior performance. As Chinese models deliver exceptional performance, will well-known models from the USA, like Meta’s Llama, be replaced in the coming years?

Source: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard

What Else Is Happening in AI on June 27th 2024❗

🏛️ Anthropic extends Claude’s access to government

Anthropic is making Claude 3 Haiku and Claude 3 Sonnet available for government use through the AWS marketplace. This will improve citizen services, streamline document reviews, support data-based policymaking, and generate realistic training scenarios.

Source: https://www.anthropic.com/news/expanding-access-to-claude-for-government?

🌐 Opera announces new themes and features for One R2

Major features include AI-based image and voice generation, image understanding capabilities, floating windows, new themes, design elements, and split window tabs.

Source: https://press.opera.com/2024/06/26/opera-unveils-opera-one-r2-in-developer

🎙️ NBC to use AI voice for Olympics recaps!

The AI will offer daily, personalized coverage of the 2024 Paris Olympics on the Peacock app in the voice of the famous sports commentator, AI Michael. According to NBC, the voice AI is trained on data from Michael’s appearances on the show.

Source: https://www.nbcnews.com/news/nbc-use-ai-generated-version-al-michaels-voice-summer-olympics-rcna159108

📹 Caption’s new feature can auto-edit videos!

The feature adds custom graphics, zooms, music, sound effects, transitions, and motion backgrounds to an unedited video. Users can also use Caption’s AI avatars to create videos and fully edit them with the AI edit feature.

Source: https://www.captions.ai/

💡 Sonia’s ‘AI Therapist’ sparks debate

The personalized AI therapist can deal with depression, stress, anxiety, poor sleep, and more by analyzing what users say during therapy sessions. The release has sparked a debate around user data privacy, its susceptibility to bias, and the accuracy of its diagnosis. Source: https://www.soniahealth.com/

🏆 Hugging Face updates Open LLM Leaderboard

Hugging Face just introduced a new upgrade to its Open LLM Leaderboard, adding new benchmarks and evaluation methods to help address the recent plateau in LLM performance gains.

  • The leaderboard now features six new benchmarks designed to be more challenging and less prone to contamination.
  • Initial rankings show Qwen2-72B-Instruct leading the pack, followed by Meta’s Llama-3-70B-Instruct and Mixtral 8×22b.
  • A new normalized scoring system adjusts for baseline performance, providing a more fair comparison across different evaluation types.
  • The upgrade also introduces a ‘maintainer’s highlight’ category and community voting system to prioritize the most relevant models.

As LLMs approach human-level performance on most tasks, finding new ways to evaluate them is becoming more difficult — and more crucial. This revamp helps guide researchers and developers towards more targeted improvements, providing a more nuanced assessment of model capabilities.

Source: https://huggingface.co/spaces/open-llm-leaderboard/blog

🚨 Rabbit R1 hit with major security flaw

A group of developers just discovered a major vulnerability in Rabbit’s R1 AI assistant device, potentially exposing user’s private data and chat responses.

  • A community-led group called Rabbitude uncovered hardcoded API keys in Rabbit’s codebase, which allowed access to all R1 responses.
  • The group gained access to the codebase in mid-May, saying the Rabbit team was aware of the issue but failed to take action.
  • Rabbitude said the vulnerability could allow bad actors to disable all r1 devices, alter voices and responses, and access private messages.
  • Rabbit acknowledged an ‘alleged data breach’ via a Discord post, but claims no customer data was leaked.

Despite massive hype in the first wave of consumer AI standalone devices, the Rabbit r1 has been nothing short of a disaster so far. Already facing major criticism over the companion’s limited capabilities, this security breach only furthers the skepticism surrounding the early AI hardware market entrants.

Source: https://rabbitu.de/articles/security-disclosure-1?

A  Daily chronicle of AI Innovations June 26th 2024:

💼 Anthropic rolls out Claude’s cutting-edge collaborative features
🤖 Google experiments with celebrity-inspired AI Chatbots
🛑 OpenAI postpones the launch of ChatGPT voice mode

🧬 New AI simulates 500M years of evolution

⏰ OpenAI delays ChatGPT voice upgrade

💼 Anthropic rolls out Claude’s cutting-edge collaborative features

Anthropic has introduced new collaboration features for Claude. These features include:

  • Projects: Projects in Claude allow integration of internal resources like style guides or codebases, enhancing Claude’s ability to deliver tailored assistance across various tasks. Users can set custom instructions for each Project to modify Claude’s tone or perspective for a specific role or industry.
  • Artifacts: It allows users to generate and edit various content types like code, documents, and graphics within a dedicated window. This benefits developers by offering larger code windows and live previews for easier front-end reviews.
  • Sharing Features: Claude Team users can share snapshots of their best conversations with Claude in their team’s shared project activity feed.

Additionally, any data or chats shared within Projects will not be used to train Anthropic’s generative models without a user’s explicit consent.

Why does it matter?

By offering advanced AI collaboration tools, Anthropic meets critical enterprise needs for flexibility and team efficiency, where many of its competitors fall short. This approach could broaden its user base, increasing user retention and market penetration.

Source: https://www.anthropic.com/news/projects

🤖 Google experiments with celebrity-inspired AI Chatbots

These chatbots will be powered by Google’s Gemini family of LLMs. The company aims to strike partnerships with influencers and celebrities and is also working on a feature that allows people to create their own chatbots by describing their personalities and appearances.

The project is led by Ryan Germick, a longtime executive at Google and a team of ten. These chatbots could be an experiment and may only appear on Google Labs rather than being widely available.

Why does it matter?

Google’s AI chatbots featuring celebrities and influencers enhance user engagement but raise ethical concerns about consent, boundaries, and legal and privacy issues as AI evolves.

Source: https://www.msn.com/en-us/news/other/google-wants-to-build-ai-chatbots-based-on-celebs-influencers-for-some-reason/ar-BB1oS1or

🛑 OpenAI postpones the launch of ChatGPT voice mode

Originally planned for late June, the Voice Mode aims to provide a more naturalistic and conversational experience with the AI chatbot, complete with emotional inflection and the ability to handle interruptions.

However, it will now be available only to a small group of users in late July or early August. OpenAI is working on improving content detection and user experience before wider rollout. GPT-4o’s real-time voice and vision capabilities are also expected to roll out to ChatGPT Plus users soon.

Why does it matter?

Delays can give competitors like Anthropic a chance to gain an edge. If rivals manage to launch similar or superior voice capabilities sooner, OpenAI risks losing potential users and market share.

Source: https://techcrunch.com/2024/06/25/openai-delays-chatgpts-new-voice-mode

🧬New AI simulates 500M years of evolution

Biotech startup EvolutionaryScale just introduced ESM3, a new AI model capable of designing and simulating new proteins in ways that mimic millions of years of natural evolution.

  • ESM3 is a frontier language model that can reason over protein sequences, structures, and functions simultaneously.
  • In tests, ESM3 generated a new fluorescent protein 58% different from other known structures — a leap that typically takes nature 500M years of evolution.
  • The largest version of ESM3 has 98B parameters, and was trained on approximately 1.8T tokens of biological data.
  • EvolutionaryScale was created by former META researchers, and has already secured $142M in seed funding from tech giants like Amazon and Nvidia.

ESM3 is bringing us closer to a future where we can ‘program’ biology as easily as we write computer code — potentially revolutionizing fields from healthcare to environmental science and unlocking completely new areas of knowledge across fields.

Source: https://www.evolutionaryscale.ai/blog/esm3-release

What Else Is Happening in AI on July 26th 2024❗

🧬 Ex-Meta engineers launch protein-generating AI biology model

EvolutionaryScale, launched by ex-Meta engineers, introduced ESM3, a gen AI model for designing novel proteins. It uses AI to simulate evolutionary processes, aiming to revolutionize industries like medicine and biofuels. It is available via API to selected customers and non-commercial users.

Source: https://venturebeat.com/ai/meta-alum-launches-ai-biology-model-that-simulates-500-million-years-of-evolution

🛑 Reddit sets new barriers for AI crawlers to protect its content

Reddit will rate-limit or block unknown bots and crawlers that violate its Public Content Policy or lack an agreement with the platform. These changes aim to deter AI companies from training LLMs on Reddit content (excluding the $60 million deal with Google).

Source: https://techcrunch.com/2024/06/25/reddits-upcoming-changes-attempt-to-safeguard-the-platform-against-ai-crawlers

🎥 AI-generated ad film debuts at Cannes 2024 

Toys ‘R’ Us, the iconic toy store chain, collaborated with OpenAI’s Sora  to produce a brand film called “The Origin of Toys ‘R’ Us.” The film premiered at the 2024 Cannes Lions International Festival of Creativity.

Source: https://venturebeat.com/ai/toys-r-us-unveils-first-commercial-made-with-openais-sora

🦊 Mozilla’s Firefox Nightly now lets users pick from leading AI chatbots

It allows users to choose from multiple AI chatbots, such as ChatGPT, Google Gemini, HuggingChat, and Le Chat Mistral. The feature lets users select their preferred AI to perform tasks such as summarizing web content, simplifying text, or conducting quizzes.

Source: https://techcrunch.com/2024/06/25/firefox-now-lets-you-choose-your-preferred-ai-chatbot-in-its-nightly-builds

📧 Google extends Gemini AI to Gmail to streamline email tasks

Google has integrated its Gemini AI into Gmail for Workspace subscribers, enhancing email functionality. It assists users by automatically drafting emails, summarizing threads, crafting responses, and retrieving information from emails or Google Drive.

Source: https://techcrunch.com/2024/06/25/google-brings-its-gemini-ai-to-gmail-via-a-sidebar-that-can-help-you-write-and-summarize-emails

A  Daily chronicle of AI Innovations June 25th 2024:

🤖 Amazon is working on a ChatGPT competitor

🎤 Google is building AI chatbots modeled after celebrities, YouTube influencers

📉 Nvidia’s market value plummets drastically

🤷‍♂️ Apple rejected Meta’s AI integration over privacy issues

🤝 OpenAI buys a remote collaboration platform

🎥 Synthesia 2.0: World’s 1st AI video communication platform
🤝 OpenAI is on an acquiring spree, buying Rocket and Multi
🎵 Record labels sue AI music startups over copyright infringement

🤖 Amazon is working on a ChatGPT competitor

  • Amazon is developing an AI chatbot, code-named “Metis,” to compete with OpenAI’s ChatGPT, and it could be launched later this year, based on a report from Business Insider.
  • The upcoming chatbot will be accessed through a web browser and rely on a new foundational model, as mentioned by sources and documents cited by Business Insider.
  • Amazon’s CEO Andy Jassy is reportedly directly involved with “Metis,” reflecting Amazon’s broader strategy to intensify its AI efforts across various applications and services.
  • Source: https://www.geekwire.com/2024/report-amazon-developing-ai-chatbot-that-would-compete-with-chatgpt-and-others/

🎤 Google is building AI chatbots modeled after celebrities, YouTube influencers

  • Google is reportedly developing new AI chatbots modeled after celebrities and YouTube influencers, aiming to create specialized interactive experiences.
  • The initiative will use Google’s Gemini family of large language models and seeks partnerships with influencers and celebrities, allowing users to design their own chatbots by describing traits and appearance.
  • The project, led by executive Ryan Germick and a team of ten, might initially be available only on Google Labs, keeping the bots experimental and not widely accessible at launch.
  • Source: https://www.engadget.com/google-is-reportedly-building-ai-chatbots-based-on-celebrities-and-influencers-235731655.html

📉 Nvidia’s market value plummets drastically

  • Nvidia’s shares dropped nearly 13% in a week, causing a loss of $430 billion from its market value and dropping its market capitalization to $2.91 trillion.
  • This decline has moved Nvidia from the top position to third place among the world’s most valuable companies, now trailing behind Microsoft and Apple.
  • Investor enthusiasm for Nvidia’s role in AI is waning after significant stock gains, leading to a 6.7% drop on Monday, the third consecutive day of declines for the company.
  • Source: https://edition.cnn.com/2024/06/25/tech/nvidia-stock-fall-worlds-third-biggest-company/index.html

🤷‍♂️ Apple rejected Meta’s AI integration over privacy issues

  • Apple declined a potential AI partnership with Meta concerning the integration of Meta’s large language model into iOS, primarily due to privacy concerns highlighted in a Bloomberg report.
  • While initial discussions between Apple and Meta took place, they did not advance, with Apple choosing not to pursue formal talks or integration due to Meta’s insufficient privacy practices.
  • Despite rejecting Meta, Apple has engaged in collaborations with other major AI entities, finalizing a deal with OpenAI to integrate ChatGPT into iOS 18 and discussing potential integrations with Google and Anthropic for diverse AI model options.
  • Source: https://www.macrumors.com/2024/06/24/apple-meta-ai-privacy-concerns/

🤝 OpenAI buys a remote collaboration platform

  • OpenAI is advancing its AI technology by integrating it into desktop computing, moving beyond browser-based interactions with the acquisition of startup Multi.
  • By purchasing Multi, OpenAI aims to develop potentially a new AI-based operating system or at least an interaction layer for current systems, as suggested by recent product releases including a ChatGPT desktop app.
  • OpenAI’s future AI assistant could replace traditional browsers, altering the role of search engines and websites, and revolutionizing how we interact with online content.
  • Source: https://the-decoder.com/openai-moves-toward-its-own-ai-operating-system-and-a-post-browser-world/

 Synthesia 2.0: World’s 1st AI video communication platform

Synthesia is launching Synthesia 2.0 – the world’s first AI video communications platform for businesses. It reinvents the entire video production process, allowing companies to create and share AI-generated videos at scale easily.

https://youtu.be/gZaBwdru_bk?si=fP7_nILeu66Oqj4B

The key new features and capabilities of Synthesia 2.0 include:

  • 2 Personal AI Avatars: Expressive Avatars shot in a studio and Custom Avatars created using your webcam.
  • AI Video Assistant: Converts text, documents, or websites into high-quality videos, with options to customize the branding, tone, and length.
  • Intuitive Video Editing: Editing simplified with “Triggers” that let you control animations and edits from the script.
  • Translation and Dynamic Video Player: Videos can now be translated into over 120 languages. Synthesia is also building a new video player with interactive features.
  • AI Safety Focus: Synthesia is pursuing ISO/IEC 42001 certification, the first standard for responsible AI management, to ensure its AI technologies are ethical.

Why does it matter?

By making video creation and distribution more accessible and automated, Synthesia aims to help businesses transition to a “video-first” approach to communicate more effectively, improve employee training, customer service, and marketing efforts, and streamline various business operations.

Source: https://www.synthesia.io/post/introducing-synthesia-video-communications-platform

🤝 OpenAI is on an acquiring spree, buying Rocket and Multi

Last week, OpenAI acquired Rockset, a startup that develops tools for real-time data search and analytics. OpenAI said it would integrate Rockset’s technology to power its infrastructure and offerings across products.

This week, OpenAI acquired Multi, a startup focused on building remote collaboration tools and software. Technically, the deal is an acqui-hire as the entire Multi team, including its co-founders, will join OpenAI to work on the company’s ChatGPT desktop application.

Why does it matter?

With over 600,000 paid business users, including most Fortune 500 companies, OpenAI’s back-to-back acquisitions strategically position it to expand its enterprise-focused AI products and make its powerful language models more useful for businesses.

Source: https://techcrunch.com/2024/06/24/openai-buys-a-remote-collaboration-platform

🎵 Record labels sue AI music startups over copyright infringement

The world’s major record labels, including Universal Music Group, Sony Music, and Warner Music, have filed twin lawsuits against the AI music generation startups Suno and Udio. The lawsuits accuse the companies of unlawfully training their AI models on massive amounts of copyrighted music, which, according to the complaints, allows the startups to generate similar-sounding music without permission.

The record labels allege Suno and Udio have effectively copied artists’ styles and specific musical characteristics. The labels claim the AI-generated music is so close to the original that it is eerily similar when transcribed into sheet music. The lawsuits also accuse the startups of making it easy for people to distribute AI-created samples that mimic copyrighted recordings on platforms like Spotify.

Why does it matter?

This legal battle is only the latest example, and one of many, of AI technology colliding with creative industries. However, any resolution of these battles will likely have far-reaching implications for the music industry and the development of AI-powered music tools.

Source: https://venturebeat.com/ai/record-labels-sue-ai-music-generator-startups-suno-udio-for-copyright-infringement

What Else Is Happening in AI on June 25th❗

🍎 Apple withdraws AI collaboration deal with Meta over privacy concerns

Contrary to the rumors about a potential AI collaboration, Apple is not planning a partnership to integrate Meta’s AI models into its products due to privacy concerns. Instead, Apple is focusing on partnerships with OpenAI and Google that align with its commitment to user privacy.

Source: https://techcrunch.com/2024/06/24/apple-shelved-the-idea-of-integrating-metas-ai-models-over-privacy-concerns-report-says

🛍️ Shopify expands its AI capabilities to attract more businesses

Shopify has expanded access to its AI-powered features, including the Sidekick assistant and image-generation tool. It is making these AI tools available to more users, as it aims to provide merchants with an integrated and seamless platform to scale their businesses.

Source: https://www.reuters.com/technology/artificial-intelligence/shopify-expands-access-its-ai-powered-features-attract-more-businesses-2024-06-24

🏫 Google is bringing Gemini access to teens through school accounts

Google now makes its AI assistant Gemini available to teenage students through school accounts. This move is aimed at helping prepare students for a future where generative AI is more prevalent. Google says Gemini will provide real-time feedback and support to students while including safeguards to prevent inappropriate responses.

Source: https://techcrunch.com/2024/06/24/google-is-bringing-gemini-access-to-teens-using-their-school-accounts

🤖 NVIDIA deploys its AI tech in the Middle East among U.S AI export curb

NVIDIA has struck a deal with Qatari telecom provider Ooredoo to deploy its AI technology in the Middle East. This move comes amid U.S. export restrictions on advanced AI chips to some countries in the region. Ooredoo will be the first company in the Middle East to offer Nvidia’s AI and graphics technology to its clients across multiple countries.

Source: https://www.reuters.com/technology/nvidia-launch-middle-east-amid-us-curbs-ai-exports-region-ooredoo-ceo-says-2024-06-23

📷 Meta marks real photos with ‘Made by AI’ label

Meta has been incorrectly labeling real photos as “Made by AI.” Multiple photographers have reported that Meta has added this label to their genuine photographs, even when they have only made minor edits using tools like Adobe’s Generative Fill. Meta says it relies on industry standards to detect AI-generated content, leading to false positives.

Source: https://www.theverge.com/2024/6/24/24184795/meta-instagram-incorrect-made-by-ai-photo-labels

A  Daily chronicle of AI Innovations June 24th 2024:

🤝 Apple and Meta are discussing gen AI partnership
🤖 ByteDance and Broadcom collaborate on AI chip development
🔍 🕵️‍♂️ Researchers developed a new method to detect hallucinations

🤝 Apple in talks with Meta for potential AI integration

🎵 Record labels sue AI music companies

🩸 AI blood test for Parkinson’s detection

Researchers just developed an AI-powered blood test that can predict Parkinson’s disease up to 7 years before symptoms appear, potentially revolutionizing early diagnosis and treatment.

  • The AI analyzes key markers in blood samples linked to inflammation and protein degradation, attempting to predict which patients would develop the disease.
  • In a 10-year study, the test correctly identified 79% of patients who later developed Parkinson’s.
  • The AI model predicted Parkinson’s up to 7 years before clinical symptoms emerged in some cases.
  • Researchers hope to develop a simple finger-prick version for widespread screenings.

AI’s ability to find subtle patterns in previously undetectable biological data is unlocking new diagnostic capabilities. By catching a debilitating disease years before the onset of symptoms, patients can get earlier access to treatment aimed at prevention.

Source: https://www.bbc.com/news/articles/c722yly6l6ro

🤝 Apple in talks with Meta for potential AI integration

  • Apple is reportedly negotiating with Meta to integrate Meta’s generative AI model into Apple’s new AI system, Apple Intelligence, according to The Wall Street Journal.
  • Apple is seeking partnerships with multiple AI companies, including Meta, to enhance its AI capabilities and catch up in the competitive AI race.
  • A potential collaboration between Apple and Meta would be significant due to their history of disagreements, and it could greatly impact the AI industry if successful.
  • Source: https://www.benzinga.com/news/24/06/39447797/apple-eyes-meta-for-potentially-game-changing-ai-partnership-report

🎵 Record labels sue AI music companies

  • Universal Music Group, Sony Music Entertainment, and Warner Records are suing AI companies Suno and Udio for massive copyright infringement related to the creation of original songs using generative AI technology.
  • The music labels claim that Suno and Udio used copyrighted works without permission, and are seeking damages of up to $150,000 per violated work in lawsuits filed in Boston and New York federal courts.
  • This legal action is part of a broader conflict between the music industry and tech companies offering AI tools, highlighting concerns about AI-generated music impacting the control and revenue of human artists.
  • Source: https://www.theverge.com/2024/6/24/24184710/riaa-ai-lawsuit-suno-udio-copyright-umg-sony-warner

🤝 Apple and Meta are discussing gen AI partnership

Apple is reportedly in talks with its longtime rival Meta to integrate the latter’s Llama 3 AI model into Apple Intelligence. This move comes as Apple prepares to roll out its AI features across iPhones, iPads, and Macs later this year.

The potential partnership follows Apple’s existing deal with OpenAI, suggesting a collaboration strategy rather than solo development in the AI race. In Apple’s arrangement with OpenAI, there’s no direct payment. Instead, OpenAI can offer premium subscriptions through Apple Intelligence, with Apple taking a percentage. It’s unclear if Meta would agree to a similar business model, given that Llama 3 is open-source and free to access.

Why does it matter?

Despite their historical rivalry, especially after Apple’s 2021 privacy changes impacted Meta’s ad revenue, this potential partnership suggests a willingness to cooperate in the AI race. However, not everyone is thrilled about it. Users and investors are voicing concerns about privacy, with Meta’s track record being a major sticking point. Some see this as a desperate move by Apple.

Source: https://www.wsj.com/tech/ai/apple-meta-have-discussed-an-ai-partnership-cc57437e

🤖 ByteDance and Broadcom collaborate on AI chip development

ByteDance is collaborating with U.S. chip designer Broadcom to develop an advanced AI processor. This partnership aims to secure a stable supply of high-end chips amid ongoing U.S.-China tensions. The project centers on creating a 5-nanometre, customized Application-Specific Integrated Chip (ASIC) that complies with U.S. export restrictions.

This chip’s manufacturing is set to be outsourced to Taiwan Semiconductor Manufacturing Company (TSMC), though production is not expected to begin this year. While the design work is currently underway, the critical “tape out” phase has yet to commence.

Why does it matter?

This partnership shows how Chinese tech companies are getting creative to secure advanced AI chips. As ByteDance has launched a major push in generative AI, this new initiative brings dual benefits: reducing procurement costs and ensuring a stable supply of higher-end chips.

Source: https://www.reuters.com/technology/artificial-intelligence/chinas-bytedance-working-with-broadcom-develop-advanced-ai-chip-sources-say-2024-06-24

🔍 🕵️‍♂️ Researchers developed a new method to detect hallucinations

ChatGPT and Gemini can produce impressive results but often “hallucinate” false or unsubstantiated information. This research focuses on a subset of hallucinations called “confabulations,” where LLMs generate answers that are both wrong and arbitrary. Researchers have developed new methods to detect confabulations using entropy-based uncertainty estimators. They introduce the concept of “semantic entropy” to measure the uncertainty of LLM generations at the meaning level.

High semantic entropy corresponds to high uncertainty and indicates a higher likelihood of confabulation. The method computes uncertainty at the level of meaning rather than specific word sequences, addressing the fact that one idea can be expressed in many ways. The method provides scalable oversight by detecting confabulations that people might otherwise find plausible.

Why does it matter?

Detecting confabulations can improve question-answering accuracy for state-of-the-art LLMs. The method helps users understand when to be cautious with LLM outputs and opens up new possibilities for using LLMs that were previously limited by unreliability.

Source: https://www.nature.com/articles/s41586-024-07421-0

What Else Is Happening in AI on June 24th 2024❗

🆕Stability AI has appointed a new CEO

Prem Akkaraju, former CEO of Weta Digital, will step up to this role. This change comes after founder Emad Mostaque resigned in March amid financial concerns. Akkaraju is part of a new investor group, including former Facebook President Sean Parker, that provides a cash infusion to Stability AI. Source: https://www.theinformation.com/articles/stability-ai-gets-new-ceo-and-a-bailout-from-investor-group-led-by-sean-parker

🖼️ Coders spotted Midjourney references in Grok’s standalone version

It hints at a possible collaboration between xAI and Midjourney for image generation. Grok might use Midjourney models via API to generate images. This could mean that the standalone Grok might let users upload images. All this could be part of the Grok 1.5 rollout. Source: https://www.bolnews.com/technology/2024/06/grok-1-5-potential-use-of-midjourney-for-image-generation-gains-evidence

🚫YouTube fights AI fakes with new policy 

YouTube expanded its privacy policy to cover AI-generated content. Now, users can request the removal of synthetic media that imitates their face or voice. However, privacy complaints alone may not guarantee content removal. While reviewing requests, YouTube will consider the potential for confusion and intent (parody/satire). Source: https://support.google.com/youtube/thread/281049221

✨ OmniAI transforms unstructured enterprise data for AI and analytics use

Founded by Y Combinator alums, OmniAI syncs with your existing storage, prep the data, and lets you run your choice of AI models on the data. It offers integrations with popular models like Meta’s Llama 3 and Anthropic’s Claude. Source: https://techcrunch.com/2024/06/22/omniai-transforms-business-data-for-ai

🤖 The new Dot AI chatbot wants to be your companion 

Dot chatbot aims to personalize interactions based on users’ thoughts and feelings. It uses multiple AI models from OpenAI and Google. The app uses intensive Q&A to learn about users’ interests and experiences. Dot offers a subscription model at $11.99/month for unlimited conversations. Source: https://techcrunch.com/2024/06/21/dots-ai-really-really-wants-to-get-to-know-you

Y Combinator joined over 140 AI startup founders in signing a letter opposing California’s Senate Bill 1047 on AI regulation, arguing it harms the ability to retain talent and stifles innovation. Source: https://www.politico.com/newsletters/california-playbook/2024/06/21/little-tech-brings-a-big-flex-to-sacramento-00164369
Formula 1 and AWS unveiled ‘Statbot’, an AI-powered system that provides real-time context, trivia, and predictions for personalized race broadcasts. Source: https://www.bloomberg.com/news/articles/2024-06-23/formula-1-and-amazon-aim-for-ai-powered-personalized-race-viewing
OpenAI announced the acquisition of Rockset, with the search and database analytics startup’s tech set to boost retrieval infrastructure for enterprise products. https://openai.com/index/openai-acquires-rockset
Axel Peytavin demonstrated a small, autonomous robot prototype powered by GPT-4o that can see, think and act based on prompts, with capabilities like picking up trash and shaking hands. Source: https://x.com/ax_pey/status/1804209628680720746
OpenAI CTO Mira Murati said in an interview that the next generation of GPT will be ‘PhD-level’ compared to GPT-3 (toddler) and GPT-4 (high school), and will be released in a year and a half. Source: https://x.com/tsarnick/status/1803901130130497952
Unitree posted new demos of its Go2 robotic dog performing handstands, flips, and withstanding and adjusting to force, showing extreme precision and agility. Source: https://x.com/UnitreeRobotics/status/1803728664213012520

A  Daily chronicle of AI Innovations June 20th 2024:

🖥️ Microsoft debuts a vision-foundational model for diverse tasks
👨‍💼 Ex-OpenAI co-founder launches own AI company
🤖 Can AI read minds? New model can predict anxiety levels 

😱 Sutskever launches new AI venture

🚧 Tech giants team up on AI factory for xAI

🚨 Perplexity faces legal action

🤖 OpenAI co-founder announces new AI company

🚀 Anthropic announces its most powerful AI yet

🇪🇺 EU Council withdraws Chat Control vote

🚧 Apple Intelligence plans face legal challenges in China

🖥️ Microsoft debuts a vision-foundational model for diverse tasks

Florence 2, Microsoft’s vision-AI model, can perform diverse tasks like object detection, captioning, visual grounding, and visual question answering via image and text prompts. It displays excellent captioning, object detection, visual grounding, and segmentation.

The model comes in 232M and 771M parameter sizes and uses a sequence-to-sequence architecture, enabling multiple vision tasks without needing a task-specific architecture modification.

On fine-tuning the model with publicly available human-annotated data, Florence 2 showcased impressive results, offering tough competition to existing large vision models like Flamingo despite its compact size.

Why does it matter?

The model will equip enterprises with a standard approach to handling various vision-handling applications. This will save resources spent on separate task-specific vision models that need fine-tuning. Moreover, it may also be useful to developers as it would eliminate the need for separate vision models for smaller tasks, significantly saving compute costs.

Source: https://www.microsoft.com/en-us/research/publication/florence-2-advancing-a-unified-representation-for-a-variety-of-vision-tasks

👨‍💼 Ex-OpenAI co-founder launches own AI company

Just a month after leaving OpenAI, ex-cofounder Ilya Sutskever has launched his own AI company, Safe Superintelligence Inc. (SSI), alongside former Y Combinator partner Daniel Gross and ex-OpenAI engineer Daniel Levy as co-founders.

According to the SSI’s launch statement on X, the company will prioritize safety, progress, and security. Sutskever also emphasizes that the company’s “singular focus” on a joint approach to safety and capabilities will prevent it from being distracted by management overhead or production cycles, unlike companies like OpenAI or Google.

Why does it matter?

SSI’s launch clearly marks the emergence of a new key player in the race to build safe, powerful AI. Its mission statement emphasizes safety and the potential for groundbreaking developments that may shape the future of AI research and development. It would be interesting to see whether the startup will uphold its mission statement in the coming days.

Source: https://ssi.inc/

🤖 Can AI read minds? New model can predict anxiety levels 

Researchers at the University of Cincinnati have developed an AI model that can identify people with an urgent risk of anxiety. The AI model uses minimal computational resources, a short picture rating task, and a small set of variables to make the prediction. The approach named “Comp Cog AI” integrates computational cognition and AI.

Participants rated 48 pictures with mildly emotional subject matter based on the degree to which they liked or disliked those pictures. The response data was then used to quantify the mathematical features of their judgments. Finally, the data was combined with ML algorithms to identify their anxiety levels.

Since the technology doesn’t rely on a native language, it is accessible to a wider audience and diverse settings to assess anxiety.

Why does it matter?

The picture rating feature can provide unbiased data to medical professionals on a person’s mental health status without subjecting them to direct questions that may trigger negative emotions. Given its 81% accuracy rate, the tool can become a useful app for detecting individuals with high anxiety risks.

Source: https://www.nature.com/articles/s44184-024-00074-x.epdf

🚀 Anthropic announces its most powerful AI yet

  • Anthropic has launched Claude 3.5 Sonnet, a new AI model that aims to be on par with, or superior to, OpenAI’s GPT-4o and Google’s Gemini across various tasks.
  • Claude 3.5 Sonnet claims to be significantly faster than its predecessor and outperforms it, even surpassing other leading models in multiple benchmarks.
  • Alongside the new model, Anthropic introduced the Artifacts feature, which allows users to interact with and edit Claude’s outputs directly within the app, enhancing its functionality beyond a typical chatbot.
  • Source: https://www.theverge.com/2024/6/20/24181961/anthropic-claude-35-sonnet-model-ai-launch

🚧 Apple Intelligence plans face legal challenges in China

  • Apple’s push to enhance its artificial intelligence capabilities faces significant challenges in China due to the country’s stringent AI regulations and censorship laws.
  • China’s strict regulation of AI, including rules about language models and data protection, means Apple must navigate approvals and potentially partner with local tech giants like Baidu and Alibaba.
  • Apple’s focus on user privacy and the need for localized AI experiences may help it gain regulatory approval but also present significant hurdles in ensuring compliance and acceptance in the Chinese market.
  • Source: https://www.cnbc.com/2024/06/20/apple-ai-push-faces-big-challenges-in-china.html

🚧 Tech giants team up on AI factory for xAI

 Dell, Nvidia, and Super Micro Computer are collaborating to build an AI factory for xAI’s massive supercomputer, which will train and scale its Grok AI chatbot.

  • Dell is assembling half of the server racks for xAI’s planned supercomputer, with Super Micro handling the other half.
  • The system aims to use up to 100,000 Nvidia H100 GPUs, potentially making it 4x larger than the biggest existing AI clusters.
  • Musk wants the supercomputer to be fully operational by Fall 2025 to accelerate the development of advanced Grok models.

This powerhouse partnership shows the staggering scale of compute power required to push the boundaries of AI. With huge capital flowing in, this massive cluster could provide a critical edge in evolving Grok to be a serious competitor to the likes of OpenAI and Anthropic.

Source: https://www.fastcompany.com/91143745/dell-super-micro-musk-xai-supercomputer-server-racks

🚨 Perplexity faces legal action

News publisher Forbes is reportedly threatening legal action against AI search startup Perplexity, alleging the company’s chatbot stole and republished original reporting without proper attribution.

  • Forbes claims Perplexity’s chatbot summarized one of its investigative articles, copying with minimal sourcing.
  • The AI-generated version was then pushed to Perplexity subscribers and turned into a YouTube video that outranked Forbes’ own content.
  • Perplexity CEO Aravind Srinivas responded that the product still has “rough edges”, and the company is working on ways to align with publishers.
  • A WIRED investigation found Perplexity scraping sites that actively blocked crawlers, while also hallucinating info and containing incorrect attributions.

Tensions between AI firms and media companies continue to escalate. While legal grounds surrounding scraping and repurposing with AI are still murky, publishers are in a bind when it comes to protecting IP. The only viable options right now seem to be cashing out with licensing deals or taking legal action.

Source: https://www.axios.com/2024/06/18/forbes-perplexity-ai-legal-action-copyright

What Else Is Happening in AI on June 20th 2024❗

🤝 Deloitte, HPE, and NVIDIA partner up! The alliance combines Deloitte’s deep industry expertise and AI capabilities with the newly released HPE NVIDIA AI Compute solutions suite. The collaboration further seeks to advance industry-specific gen AI applications and assist clients across sectors in modernizing their data strategies by accessing innovative insights.

Source: https://www.prnewswire.com/news-releases/deloitte-announces-collaboration-with-hewlett-packard-enterprise-and-nvidia-on-co-developed-generative-ai-solutions-302175762.html

🚗 NVIDIA’s AI can turbocharge deployment of self-driving cars: NVIDIA has unveiled a new AI software, Omniverse Cloud Sensor RTX, that will likely accelerate the development of self-driving cars and robots.

The software combines real-world data with synthetic data, making it easy to test sensor perception in realistic virtual environments before deployment into the actual world.

Source: https://nvidianews.nvidia.com/news/omniverse-microservices-physical-ai

🎶 YouTube trials AI-powered “Ask for music” feature: The feature seeks to make searching music more conversational for users. Users can search for music using voice commands and verbal prompts. It is speculated that the feature may be an addition to the AI-generated playlist cover available on YouTube.

Source: https://www.business-standard.com/topic/youtube

🎥 Luma adds an “extend video” feature to Dream Machine: In contrast to Luma’s previous five-second limit for videos, the “extend video” feature will allow videos to be extended based on prompts. The AI model will consider the new context while extending the video. Additionally, the upgrade will enable Standard, Pro, and Premier users to remove watermarks.

Source: https://x.com/LumaLabsAI/status/1803098863097835995

🛠️ Apple releases 20 core ML models on Hugging Face: The release includes 4 major datasets in addition to 20 core models as a part of Apple’s efforts to equip developers with advanced on-device AI capabilities.

These core ML models have been optimized to run exclusively on users’ devices and can be used for various applications like image classification, depth estimation, and semantic segmentation.

Source: https://huggingface.co/apple

A  Daily chronicle of AI Innovations June 19th 2024:

📈 Nvidia is now the world’s most valuable company

🤖 Perplexity AI is lying about their user agent

👀 Meta releases new AI models for text, image and audio

🦋 Ex-Snap engineer launches social network where AIs and humans coexist

🩺 OpenAI and Color Health join forces to accelerate cancer treatment
🎤 Meta unveils new AI models for audio, text, and watermarking
⚒️Notion introduces AI Connectors for Slack & Google Drive integration

🎺 TikTok brings AI to content creation

📈 Nvidia is now the world’s most valuable company

  • Nvidia surpassed Microsoft to become the world’s most valuable company, reaching a market cap of $3.34 trillion, driven by its focus on artificial intelligence and expanded robotics research.
  • The company’s stock price increased by 3.7%, putting it ahead of Microsoft’s $3.33 trillion market cap, while Apple ranks third with a market cap of $3.28 trillion.
  • Nvidia’s rise is fueled by partnerships and advancements in AI and robotics, including a new collaboration with Hewlett Packard to launch NVIDIA AI Computing and Private Cloud AI, along with significant projects with major AI developers like Google, Meta, Microsoft, and OpenAI.
  • Source: https://decrypt.co/235988/nvidia-flips-microsoft-top-tech-company

🤖 Perplexity AI is lying about their user agent

  • WIRED and developer Robb Knight found that Perplexity’s crawler ignores the Robots Exclusion Protocol, accessing content from sites that have blocked it, including Condé Nast properties.
  • Perplexity CEO Aravind Srinivas describes the product as an “answer engine,” but the chatbot has been found to paraphrase articles without proper attribution and sometimes fabricate information entirely.
  • An experiment by WIRED demonstrated that Perplexity sometimes generates content without accessing the actual webpage, raising concerns about the accuracy and ethics of its information-gathering methods.
  • Source: https://www.wired.com/story/perplexity-is-a-bullshit-machine/

👀 Meta releases new AI models for text, image and audio 

  • Meta’s Fundamental AI Research team introduced new models such as Chameleon, a multimodal model for text and images, and JASCO, a text-to-music model accepting various inputs like chords or beats.
  • Meta’s Chameleon can handle and generate text and images, with its 7B and 34B variants available for research under a non-commercial license.
  • The company launched AudioSeal, an audio watermarking technology, which can identify AI-generated speech, released under a commercial license for recognizing speech in lengthy audio segments.
  • Source: https://the-decoder.com/meta-releases-new-ai-models-for-text-image-and-audio/

🦋 Ex-Snap engineer launches social network where AIs and humans coexist

  • Butterflies, founded by former Snap engineer Vu Tran, is a social network allowing humans and AI personas (called Butterflies) to interact through posts, comments, and direct messages, launching publicly on iOS and Android after five months in beta.
  • Users can easily create Butterfly personas with backstories, emotions, and opinions, which then autonomously generate content that other users and AIs can engage with, aiming to provide a more immersive AI experience compared to traditional chatbots.
  • The app received $4.8 million in seed funding led by Coatue and has mostly been used for entertainment, with potential plans for subscription models and brand interactions in the future.
  • Source: https://techcrunch.com/2024/06/18/former-snap-engineer-launches-butterflies-a-social-network-where-ais-and-humans-coexist/

🩺 OpenAI and Color Health join forces to accelerate cancer treatment

In collaboration with OpenAI, Color Health has developed a copilot application that leverages OpenAI’s GPT-4.o model to assist clinicians in accelerating cancer treatment. Here’s how it works:

  • Integration of medical data: The copilot integrates patient medical data with clinical knowledge using OpenAI’s APIs. It processes and normalizes vast unstructured data such as family history, individual risk factors, and clinical guidelines.
  • Tailored treatment plans: Based on this data, the copilot generates customized, comprehensive treatment plans. These plans answer critical questions like “What screenings should the patient be doing?” and identify missing diagnostics.
  • Clinician-in-the-loop workflow: A clinician evaluates the copilot’s output at each step. They can modify it if necessary before presenting it to the patient.
  • Assured privacy: Color Health ensures HIPAA compliance and maintains rigorous privacy standards while providing clinician oversight to ensure high-quality care.

Why does it matter? 

The tool identifies missing diagnostics and expedites the analysis of complex medical records – a process that can now be completed in just 5 minutes rather than hours or weeks. This not only improves access to critical expertise but also has the potential to catch cancer or pre-cancerous conditions earlier, enabling faster treatment and better patient outcomes.

Source: https://openai.com/index/color-health

🎤 Meta unveils new AI models for audio, text, and watermarking

Meta’s Fundamental AI Research (FAIR) team has recently released several new AI models and tools for researchers to use. Here are the highlights:

  • JASCO: This AI model, short for “Joint Audio and Symbolic Conditioning for Temporally Controlled Text-to-Music Generation,” allows users to adjust features of generated sound (such as chords, drums, and melodies) through text inputs. FAIR plans to release the JASCO inference code under an MIT license and the pre-trained model under a non-commercial Creative Commons license.
  • AudioSeal: This tool adds watermarks to AI-generated speech. It’s designed specifically for localized detection of AI-generated segments within longer audio snippets and will be released with a commercial license.

Chameleon: FAIR will release two sizes of its multimodal text model, Chameleon (7B and 34B), under a research-only license. These models can handle tasks requiring visual and textual understanding, such as image captioning.

Why does it matter?

By making these tools publicly available, Meta encourages collaboration within the AI community, which could lead to groundbreaking applications in various fields, such as music creation, sound design, and content generation. Moreover, the introduction of AudioSeal addresses the growing need to identify and attribute AI-generated content, promoting transparency and accountability in an increasingly AI-driven world.

Source: https://venturebeat.com/ai/meta-releases-flurry-of-new-ai-models-for-audio-text-and-watermarking

⚒️Notion introduces AI Connectors for Slack & Google Drive integration

With AI Connectors, users can query Notion and receive responses from connected apps. The integration allows for selecting specific public channels or all public channels for information retrieval, excluding private channels and direct messages.

Notion AI can access up to a year of historical Slack content, which may take up to 36 hours. Additionally, users can set up various interactions between Notion and Slack, such as sending Slack messages to a Notion database, creating Notion tasks directly from Slack, and receiving notifications in Slack for specific events in Notion

Why does it matter?

This development demonstrates the growing trend of incorporating AI capabilities into mainstream productivity and collaboration platforms. The ability to intelligently pull in and surface relevant information from external platforms helps streamline workflows, enhance productivity, and improve the overall efficiency of the team.

Source: https://x.com/NotionHQ/status/1803096151526137934

🧠 AI model predicts anxiety levels

Researchers at the University of Cincinnati just developed a new AI system that can predict anxiety levels using a short picture rating task and a small set of demography and psychological variables.

  • Comp Cog AI combines principles from psychology with machine learning to measure human decision-making patterns related to anxiety.
  • Participants rate a series of emotion-evoking pictures and answer a few contextual questions about their age, income, employment, etc.
  • The model extracts 15 ‘judgment variables’ that capture patterns in how users weigh positive and negative outcomes in decision-making.
  • In tests, the model predicted whether participants had higher or lower anxiety with up to 81% accuracy.

While anxiety disorders are extremely prevalent, they’re also often challenging and time-consuming to diagnose. With AI’s ability to connect data patterns and achieve reliable predictions, the tech could soon enable more efficient screenings to streamline assessment and treatment time.

Source: https://www.nature.com/articles/s44184-024-00074-x.epdf

What Else Is Happening in AI on June 19th 2024❗

🧠 Google transitions DeepMind from research to AI product development

This transition involves merging Google Research’s Brain team with DeepMind, creating a unified group called Google DeepMind. The aim is to enhance the development of advanced AI technologies responsibly. However, this move has led to frustration among some researchers due to new guidelines and the pressure to commercialize AI technologies.

Source: https://www.pymnts.com/google/2024/google-reportedly-moving-deepmind-from-research-to-ai-products

📽️ ElevenLabs’ new open-source AI tool adds sound effects to videos

The app extracts four frames from the video and sends them with a prompt to OpenAI’s GPT-4 to generate a text-to-sound effects prompt. This prompt is used by ElevenLabs’s Sound Effects API to create the sound effect, which is then combined with the video into a single file up to 22 seconds long, ready for download.

Source: https://venturebeat.com/ai/elevenlabs-unveils-open-source-creator-tool-for-adding-sound-effects-to-videos

🤖 Snap unveils advanced gen AI for enhanced AR effects

Snap’s latest iteration of gen AI allows users to see more realistic special effects. It plans to create full-body AR experiences rather than just facial AR experiences, such as generating a new outfit. It also announced an upgraded version of its developer program, Lens Studio, which lets artists and developers create AR features for Snapchat or other websites and apps.

Source: https://www.reuters.com/technology/snap-launches-ai-tools-advanced-augmented-reality-2024-06-18

🎨 Hedra Labs launches Character-1, a multimodal creation studio preview

Hedra Labs has launched a preview of Character-1, a model for creating expressive characters that talk, sing, and rap. It offers infinite duration (30s in open preview) and generates 90s content per 60s of input, hardware permitting. This is part of Hedra’s mission to give creators control over dialogue, movement, and virtual worlds.

Source: https://x.com/hedra_labs/status/1803095713112580475

🦋 Ex-Snap engineer launches AI-human social network, Butterflies

The platform allows users to create AI avatars, called “Butterflies,” that can engage in conversations, generate images, and participate in social activities like human users. The app offers a range of features, including creating and customizing AI characters and exploring a feed filled with AI-generated and human-generated content.

Source: https://techcrunch.com/2024/06/18/former-snap-engineer-launches-butterflies-a-social-network-where-ais-and-humans-coexist

🎺 TikTok brings AI to content creation

TikTok just launched Symphony, a new suite of AI features to help brands and creators produce content more efficiently on the platform — including digital avatars, translation tools, an AI assistant, and more.

  • Symphony Digital Avatars enable brands to generate AI spokespeople to star in their TikTok ads and branded content across multiple languages.
  • Brands can choose from a selection of ‘stock avatars’ based on real actors or create custom avatars to serve as virtual brand reps.
  • A new AI Dubbing tool automatically transcribes, translates, and dubs videos into 10+ languages, helping brands scale content globally.
  • TikTok’s Creative Assistant has been rebranded to Symphony Assistant, a chatbot to help with brainstorming, writing scripts, and optimizing content for the platform.

The creator economy is about to be completely remade with AI, and these tools will unlock new opportunities for optimized content, reach, and creation — but the rise of digital ‘spokespeople’ will also raise new questions on authenticity and consumer trust that brands will need to grapple with.

Source: https://www.tiktok.com/business/en-US/blog/tiktok-symphony-ai-creative-suite

AI Jobs on June 9th 2024:

  • 🎉 Anthropic – Head of Events: https://jobs.therundown.ai/jobs/63479075-head-of-events
  • 📚 Tempus – Analyst, NEXT Strategy & Operations: https://jobs.therundown.ai/jobs/63500064-analyst-next-strategy-operations
  • 🎨 Fiddler AI – Staff UI/UX Designer: https://jobs.therundown.ai/jobs/63473287-staff-ui-ux-designer
  • 👥 Glean – GTM Recruiter: https://jobs.therundown.ai/jobs/63498594-gtm-recruiter

A  Daily chronicle of AI Innovations June 18th 2024:

🩻 OpenAI expands into healthcare with AI tool for cancer screening

🛰️ NATO invests $1.1 billion in AI, space, and robotics defence tech

🍎 Apple halts work on Vision Pro

🎬 Google DeepMind’s new AI can generate soundtracks for videos
🌟 Runway launches new model Gen-3 Alpha
🚀China’s DeepSeek Coder V2 beats GPT-4 Turbo

🔊 DeepMind creates sound for videos

🍎 Apple halts work on Vision Pro

  • Apple has paused development on the second-generation Vision Pro headset to concentrate on a more affordable version, according to The Information.
  • The company reportedly began working on this cheaper Vision device, codenamed “N109,” in 2022, aiming to sell it for a price similar to a high-end iPhone.
  • Even though Apple plans to use the high-end display components from the Vision Pro in the budget model, it still faces challenges in reducing costs without removing too many features.
  • Source: https://www.macrumors.com/2024/06/18/apple-suspends-work-on-vision-pro-2/

🩻 OpenAI expands into healthcare with AI tool for cancer screening

  • OpenAI has partnered with Color Health to integrate artificial intelligence into cancer screening and treatment procedures using their GPT-4o model.
  • The AI assistant developed by Color Health processes patient data to create personalized cancer screening strategies and informs doctors of missing diagnostic tests.
  • The AI tool also helps in assembling a cancer pretreatment work-up, expediting the process that can take weeks or months, and thus potentially reducing delays in treatment.
  • Source: https://www.newsbytesapp.com/news/science/openai-partners-with-color-health-for-cancer-copilot/story

🛰️ NATO invests $1.1 billion in AI, space, and robotics defence tech

  • NATO has started its €1 billion investment in defence technology, focusing on AI, space, and robotics, by funding four European tech firms and four venture capital funds.
  • The NATO Innovation Fund, established in 2022, aims to address contemporary defence and security challenges by matching government buyers with start-ups to develop new technologies.
  • Initial recipients of the fund include Germany’s ARX Robotics, UK’s Fractile AI, iComat, and Space Forge, with the broader aim of promoting NATO’s technological sovereignty and innovation ecosystems.
  • Source: https://www.euronews.com/next/2024/06/18/nato-begins-its-1-billion-investment-in-ai-space-and-robotics-defence-tech

🎬 Google DeepMind’s new AI can generate soundtracks for videos

DeepMind is developing video-to-audio (V2A) technology to generate rich soundtracks for silent videos generated by AI models. V2A combines video pixels with natural language text prompts to create synchronized audiovisual content. The technology offers enhanced creative control, allowing users to guide the audio output using positive and negative prompts.

What sets DeepMind’s V2A apart is its ability to understand raw pixels and generate audio without manual alignment. However, V2A struggles with artifacts or distortions in videos and generates audio that is not super convincing. As DeepMind continues to gather feedback from creators and filmmakers, they remain committed to developing this technology responsibly.

Why does it matter?

The technology could help revive and enhance historical footage, silent films, and other archival material. However, generative AI tools like V2A also threaten to disrupt the film and TV industry, potentially eliminating jobs without strong labor protections.

Source: https://deepmind.google/discover/blog/generating-audio-for-video

🌟 Runway launches new model Gen-3 Alpha

Runway launched Gen-3 Alpha, its latest AI model for generating video clips from text descriptions and still images. Gen-3 Alpha excels at generating expressive human characters with a wide range of actions, gestures, and emotions and can interpret various styles and cinematic terminology. However, it has limitations, including a maximum video length of 10 seconds, and struggles with complex character and object interactions and following the laws of physics precisely.

Runway partnered with entertainment and media organizations to create custom versions of Gen-3 for more stylistically controlled and consistent characters, targeting specific artistic and narrative requirements. They also have implemented safeguards, such as a moderation system to block attempts to generate videos from copyrighted images and a provenance system to identify videos coming from Gen-3.

Why does it matter?

As competition in AI video generation heats up, Runway’s Gen-3 Alpha empowers artists and filmmakers to create high-quality, controllable videos with ease, pushing the boundaries of storytelling and creative possibilities.

Source: https://runwayml.com/blog/introducing-gen-3-alpha

🚀China’s DeepSeek Coder V2 beats GPT-4 Turbo

Chinese AI startup DeepSeek has announced the release of DeepSeek Coder V2, an open-source code language model. It is built upon the DeepSeek-V2 MoE model and excels at coding and math tasks, supporting over 300 programming languages. It outperforms state-of-the-art closed-source models like GPT-4 Turbo, Claude 3 Opus, and Gemini 1.5 Pro, making it the first open-source model to achieve this feat. DeepSeek Coder V2 also maintains comparable performance in general reasoning and language capabilities.

The model is being offered under an MIT license, which allows for research and unrestricted commercial use. It can be downloaded or accessed via API on DeepSeek’s platform.

Why does it matter?

DeepSeek aims to “unravel the mystery of AGI with curiosity” and has quickly emerged as a notable Chinese player in the AI race. As it only costs $0.14/1M tokens(input) and $0.28/1M tokens(output), it will give notable models like GPT-4 Turbo intense competition.

Source: https://github.com/deepseek-ai/DeepSeek-Coder-V2/blob/main/paper.pdf

🔊 New AI Jobs Opportunities on June 18th 2024

  • 🎨 Glean – UX Designer: https://jobs.therundown.ai/jobs/63229234-ux-designer
  • 🔬 Fiddler AI – Staff AI Scientist: https://jobs.therundown.ai/jobs/63187617-staff-ai-scientist
  • 📋 Notable – Product Operations Manager: https://jobs.therundown.ai/jobs/63204832-product-operations-manager
  • 📊 Findem – Data Analyst: https://jobs.therundown.ai/jobs/63102512-data-analyst

What Else Is Happening in AI on June 18th 2024❗

🔍 Perplexity now displays weather, currency conversion, and simple math directly through cards

This move aims to keep users from going to Google for such results. Perplexity’s CEO, Aravind Srinivas, acknowledged that Google handles basic queries like weather, time, and live sports scores well, and his company had work to do in that area.

Source: https://techcrunch.com/2024/06/17/perplexity-now-displays-results-for-temperature-currency-conversion-and-simple-math-so-you-dont-have-to-use-google

🛡️ U.S. government and private sector ran the first AI attack simulation

Federal officials, AI model operators, and cybersecurity companies ran the first joint simulation of a cyberattack on a critical AI system. It also involved experts from private sector companies like Microsoft, Nvidia, and OpenAI. It helped identify potential new threats and establish communication channels between the government and the private sector.

Source: https://www.axios.com/2024/06/17/cisa-tech-companies-ai-cyberattack-simulation

🚀 Adobe Acrobat got a major upgrade, bringing AI to PDFs and more

Adobe Firefly generative AI enables image generation and editing directly within Acrobat for the first time. Acrobat AI Assistant’s new features, “insights across documents” and “enhanced meeting transcripts,” help users extract insights and share information from various document types. Adobe is offering free, unlimited access to Acrobat AI Assistant from June 18 to June 28.

Source: https://blog.adobe.com/en/publish/2024/06/17/adobe-acrobat-reimagines-documents-multi-format-ai-powered-work

🤖 TikTok introduces gen AI avatars of creators and stock actors for ads

“Custom Avatars” allow creators to scale their likeness for multilingual avatars and brand collaborations, while brands can use pre-built “Stock Avatars” to add a human touch. Plus, the new “AI Dubbing” tool translates content into ten languages, helping creators and brands increase their global reach.

Source: https://newsroom.tiktok.com/en-us/announcing-symphony-avatars

🧱 Pixelbot 3000 builds Lego art using simple AI prompts

YouTuber Creative Mindstorms designed and built the Pixelbot 3000, a Lego printer that automates the assembly of brick-built mosaics. It uses OpenAI’s DALL-E 3 to generate images based on simple text prompts. First it generates a simplified cartoon-style image, then it is divided into a 32 x 32 grid, and the color of the center pixel in each square is sampled to create a high-contrast scaled image for the mosaic.

Source: https://www.theverge.com/2024/6/17/24180250/lego-printer-pixel-art-artificial-intelligence

Adobe integrated new Firefly AI capabilities into Acrobat, allowing users to create and edit images within PDFs using text prompts — also adding the ability to access an AI assistant for insights, content creation, and more. https://news.adobe.com/news/news-details/2024/Adobe-Reimagines-Acrobat-Bringing-Firefly-AI-to-PDFs-and-Expanding-Use-Across-More-Document-Types/default.aspx
Reuters Institute of Journalism published a new report finding growing public wariness of AI-generated news content, with many expressing discomfort about its potential impact on content reliability and trust. https://www.reuters.com/technology/artificial-intelligence/global-audiences-suspicious-ai-powered-newsrooms-report-finds-2024-06-16
The U.S. Navy is deploying AI-powered underwater drones to better detect threats, with plans to expand the tech’s use in identifying enemy ships and aircrafts. https://finance.yahoo.com/news/ai-equipped-underwater-drones-helping-153947268.html
Luma teased new control features coming to its Dream Machine video model, including the ability to quickly change scenes and precisely edit characters — also launching the ability to extend video and remove watermarks. https://x.com/LumaLabsAI/status/1802678788490494128
Anthropic published new research showing that AI models can engage in ‘reward tampering’, learning to cheat the system and grant higher rewards even without specific training. https://www.anthropic.com/research/reward-tampering

A  Daily chronicle of AI Innovations June 17th 2024:

🍔 McDonald’s pauses AI drive-thru tests

🤖 Chinese startup creates realistic robots

🤷‍♂️ Photographer wins AI-image contest with real picture, gets disqualified

💻 NVIDIA’s AI model for synthetic data generation rivals GPT-4
⚠️ Meta pauses AI model training in EU due to regulatory pushback
🎵 Spotify launches ‘Creative Labs’ to test Gen AI voiceover ads

🏅 Tiny LLM matches GPT-4 on math

🍔 McDonald’s pauses AI drive-thru tests

  • McDonald’s is ending its artificial intelligence drive-thru order-taking program and will shut it off in all participating restaurants by July 26, 2024.
  • The decision to end the AI program came after a thorough review, and McDonald’s Chief Restaurant Officer stated the company will seek broader voice ordering solutions.
  • Although the AI drive-thru program showed successes, McDonald’s aims to simplify operations and speed up service, with plans to decide on a new voice ordering solution by the end of the year.
  • Source: https://www.foxbusiness.com/technology/mcdonalds-ending-ai-drive-thru-ordering

🤷‍♂️ Photographer wins AI-image contest with real picture, gets disqualified

  • A photographer named Miles Astray won an AI-image competition with a real photograph of a head-scratching flamingo, resulting in his disqualification.
  • His entry, titled F L A M I N G O N E, was mistakenly awarded third place by judges from notable institutions like the New York Times and Christie’s.
  • Astray aimed to demonstrate that human creativity and natural beauty surpass AI-generated content, sparking a discussion about the ethical implications of using AI in art.
  • Source: https://www.techspot.com/news/103422-photographer-wins-ai-image-contest-real-picture-gets.html

NVIDIA’s AI model for synthetic data generation rivals GPT-4

NVDIAI has released Nemotron-4 340B, an open-source pipeline for generating high-quality synthetic data. It includes a base model trained on 9M tokens, an instruction, and a reward model.

  • The instruction model can generate diverse synthetic data that mimics real-world data.
  • The reward model then evaluates the generated data to filter out high-quality responses.
  • This interaction between the two models produces better training data over time.

Note: 98% of the training data used to fine-tune the Instruct model is synthetic and was created using NVIDIA’s pipeline.

In benchmarks such as MT-Bench, MMLU, GSM8K, HumanEval, and IFEval, the Instruct model generally performs better than other open-source models such as Llama-3-70B-Instruct, Mixtral-8x22B-Instruct-v0.1, and Qwen-2-72B-Instruct, and in some tests, it even outperforms GPT-4o.

It also performs comparable to or better than OpenAI’s GPT-4-1106 in human evaluation for various text tasks, such as summaries and brainstorming. The technical report provides detailed benchmarks.

Why does it matter?

This development allows businesses to create powerful, domain-specific LLMs without the need for extensive, costly real-world datasets. It has significant potential impacts across various industries, such as healthcare (drug discovery, personalized medicine, medical imaging), finance (fraud detection, risk assessment, customer service), manufacturing (predictive maintenance, supply chain optimization), and retail (personalized customer experiences).

Source: https://blogs.nvidia.com/blog/nemotron-4-synthetic-data-generation-llm-training

⚠️ Meta pauses AI model training in EU due to regulatory pushback

In response to the regulatory pressure from the Irish Data Protection Commission and the UK’s Information Commissioner’s Office, Meta has decided to pause its plans to train its large language model, Llama, using public content shared by Facebook and Instagram users in the European Union and the UK.

The regulators expressed concerns about Meta’s plan to use this user-generated content to train its AI systems without obtaining explicit user consent. Meta relied on a GDPR provision called “legitimate interests” to justify this data usage, but the regulators felt this was insufficient. Meta has decided to delay the launch of its AI chatbot in Europe until it can address the regulators’ concerns and establish a more transparent user consent process.

Why does it matter?

Meta’s inability to use EU user data for AI training is a setback for its regional AI ambitions. It could disadvantage Meta against competitors who can leverage such data. This situation highlights the ongoing tensions between tech companies’ desire to utilize consumer data for AI development and regulators’ efforts to protect user privacy. Striking the right balance between innovation and privacy will be a major challenge as the AI race intensifies.

Source: https://www.bloomberg.com/news/articles/2024-06-14/meta-delays-ai-chatbot-launch-in-europe-after-regulator-pushback

🎵 Spotify launches ‘Creative Labs’ to test Gen AI voiceover ads

Spotify has launched a new in-house creative agency called “Creative Lab.” This agency will help brands and advertisers create custom campaigns for Spotify’s platform. Creative Lab teams in different markets will provide local insights and collaborate with brands to develop campaigns through workshops, inspiration sessions, and collaborative ideation.

In addition, Spotify is also testing a new AI tool called “Quick Audio” that will allow brands to create scripts and voiceovers using generative AI technology. This new capability will be integrated into Spotify’s ad manager platform, giving advertisers more options to produce audio ads for Spotify’s audience of over 615 million listeners.

Why does it matter?

This move emphasizes Spotify’s ambition to become a full-service advertising platform. Marketers and advertisers will have new creative and production capabilities available through Spotify to better reach the platform’s large and engaged user base in unique ways, including potentially using AI-generated audio ads. This could disrupt traditional advertising models and open new possibilities for how brands connect with consumers on audio platforms.

Source: https://techcrunch.com/2024/06/13/spotify-creative-labs-ad-agency-for-advertisers

🤖 Chinese startup creates realistic robots

A Chinese startup called Ex-Robots is developing ultra-realistic humanoid robots that can closely mimic human facial expressions and emotions.

  • Ex-Robots’ lifelike humanoid robots can smile, frown, and replicate subtle facial movements using small motors embedded in the silicone faces.
  • The startup uses multimodal AI and algorithms to enable the robots to recognize and express emotions based on perceiving their environment.
  • Creating one humanoid robot currently takes 2-4 weeks, with costs as high as $275k per unit.
  • The company said near-term use cases include museum exhibits, with eventual roles in healthcare, therapy, and child services.

The advancement of ultra-realistic robots combined with the power of AI is about to create some bizarre new frontiers. Will humans be creeped out by an actual human-looking humanoid over mechanical options like Tesla’s Optimus? It’s unclear if the world is ready… but we’ll find out soon.

Source: https://www.reuters.com/technology/chinas-ex-robots-develops-humanoids-with-enhanced-facial-movement-2024-06-11

🏅Tiny LLM matches GPT-4 on math

Shanghai AI Laboratory researchers just developed a new algorithm called MCT Self-Refine, which allowed a relatively small 8B-parameter model to achieve GPT-4 level performance on complex math.

  • MCT Self-Refine combines an 8B LLaMa model with Monte Carlo Tree Search, an AI technique famously used by Google DeepMind to master the game of Go.
  • The algorithm constructs a search tree, refining answers through a series of processes and self-evaluation.
  • On the GSM-Hard math word problem benchmark, infusing the Monte Carlo Tree Search took the 8B model from 25.5% accuracy to 45.5%.
  • The model also achieved 96.7% on the GSM8K benchmark, outperforming significantly larger models like GPT-4, Claude, and Gemini.

 The research shows it’s possible to boost math capabilities without massive scale — and GPT-4 level performance with a model trained on 200x less parameters is an impressive feat. If the approach proves to be a more efficient path to advanced reasoning, we could be on the cusp of a new wave of model acceleration.

Source: https://arxiv.org/pdf/2406.07394

New AI Jobs on June 17th 2024

  • 🎨 Abridge – Senior Product Designer: https://jobs.therundown.ai/jobs/63255899-senior-product-designer
  • 💻 Palantir Technologies – Software Engineer, New Grad: https://jobs.therundown.ai/jobs/63248966-software-engineer-new-grad-us-government
  • 🧪 Fiddler AI – Staff AI Scientist: https://jobs.therundown.ai/jobs/63187617-staff-ai-scientist
  • 🤖 Meta – Software Engineer – ML Systems: https://jobs.therundown.ai/jobs/63275347-software-engineer-ml-systems-technical-leadership

 

What Else Is Happening in AI on June 17th 2024❗

🍎 Apple enters the AI icon race to find a logo that makes sense

Apple has joined other tech giants like Google, OpenAI, Anthropic, and Meta in the race to find an iconic visual representation for AI. No company has yet created an unambiguous “AI logo” that conveys the concept to users. AI’s lack of a clear visual identity reflects the difficulty of representing such a broad and evolving technology in a simple icon.

Source: https://techcrunch.com/2024/06/15/apple-joins-the-race-to-find-an-ai-icon-that-makes-sense

📝 Niloom.AI launches gen AI content creation platform for spatial computing

Without extensive technical expertise, the platform allows users to create, prototype, edit, and instantly publish sophisticated AR/VR content using text or speech prompts. It consolidates the entire creative process, from ideation to publishing, and integrates with various third-party tools to provide a one-stop solution for spatial computing content creation.

Source: https://venturebeat.com/ai/niloom-ai-launches-one-stop-generative-ai-content-creation-platfor-for-spatial-computing

🏟️ AI to delete abusive posts against athletes during the 2024 Paris Olympics

The International Olympic Committee (IOC) will deploy AI at the 2024 Paris Olympics to automatically detect and erase abusive social media posts directed at athletes and officials. The AI tool will monitor posts about 15,000 athletes and officials and immediately remove any content involving hate speech, bullying, or political attacks.

Source: https://www.reuters.com/sports/olympics/ai-erase-abusive-posts-athletes-paris-olympics-ioc-2024-06-14

🖼️ Picsart and Getty team up to counter Adobe’s “commercially-safe” AI

Picsart has partnered with Getty Images to develop a “responsible, commercially-safe” AI image generator tool. The AI model will be trained exclusively on Getty’s licensed stock content to address concerns about AI-generated content violating copyright laws. Picsart hopes to provide a viable alternative to Adobe’s Firefly by leveraging Getty’s library of licensed images. Source: https://www.theverge.com/2024/6/14/24178333/picsart-getty-commercially-safe-ai-image-tool-adobe

📰 Yahoo News gets an AI-powered revamp with Artifacts integration 

Yahoo has acquired the technology behind the Artifact news aggregation app and is launching a new AI-powered Yahoo News app. The app will feature a personalized news feed based on user interests and a “Key Takeaways” feature that provides bullet-point summaries of articles. Users can also flag problematic content, which the AI will then try to rewrite.

Source: https://www.theverge.com/2024/6/13/24177980/yahoo-news-app-launch-artifact-ai-architecture

A  Daily chronicle of AI Innovations June 14th 2024:

📊 Google’s PH-LLM reads your wearables’ data for personalized insights
🔮 Ex-OpenAI researcher on what to expect from AI in next decade
🧠 DeepMind built ‘a virtual rodent’ with AI to understand brain activity

🗳️ AI chatbots run for office

💼LinkedIn is rolling out new AI-powered features for premium users

🌍Synthflow’s AI voice assistants are now multilingual!

🖼️Picsart is partnering with Getty Images to develop a custom model for AI imagery

🕵️‍♂️Former head of NSA joins OpenAI’s Safety and Security Committee

📊 Google’s PH-LLM reads your wearables’ data for personalized insights

Building on the next-gen capabilities of Gemini models, Google has presented research that highlights two complementary approaches to providing accurate personal health and wellness information with LLMs.

The first introduces PH-LLM, a version of Gemini fine-tuned to understand and reason on time-series personal health data from wearables such as smartwatches and heart rate monitors. The model answered questions and made predictions noticeably better than experts with years of experience in the health and fitness fields.

In the second paper, Google introduces an agent system that leverages state-of-the-art code generation and information retrieval tools to analyze and interpret behavioral health data from wearables. Combining these two ideas will be critical for developing truly personalized health assistants.

Why does it matter?

Wearables generate a wealth of personal health data that is rarely utilized in clinical settings. Integrating this data with advanced AI models could revolutionize personal health management and preventative care by putting an “expert health assistant” on everyone’s wrist.

Source: https://research.google/blog/advancing-personal-health-and-wellness-insights-with-ai

🔮 Ex-OpenAI researcher on what to expect from AI in next decade

A researcher fired from OpenAI, Leopold Aschenbrenner,  published a 165-page essay on what to expect from AI in the next decade. And GPT-4 has summarized it! Here are some key takeaways from the essay:

  • By 2027, AI models could reach the capabilities of human AI researchers and engineers, potentially leading to AI surpassing human intelligence
  • Trillions of dollars are being invested into developing the infrastructure needed to support these AI systems
  • Controlling AI systems smarter than humans(the ‘superalignment‘ problem) will be crucial to prevent catastrophic outcomes
  • Only a few hundred people truly understand the scale of change AI is about to bring

Why does it matter?

The essay provides a rare insider’s perspective on the rapid progression of AI. Coming from someone deeply involved in cutting-edge AI development, the insights highlight the urgency to get ahead of managing risks before AI’s capabilities outpace our defenses.

Source: https://ca.news.yahoo.com/researcher-fired-openai-published-165-184227878.html

🧠 DeepMind built ‘a virtual rodent’ with AI to understand brain activity

Researchers from Google DeepMind and Harvard built a ‘virtual rodent’ powered by AI to help them better understand how the brain controls movement. With deep reinforcement learning (RL), it learned to operate a biomechanically accurate rat model, allowing researchers to compare real and virtual neural activity.

Why does it matter?

Understanding how the brain controls movement and modeling neural activity could exponentially advance fields like neuroscience and brain-computer interfaces, with the help of AI.

Source: https://www.nature.com/articles/s41586-024-07633-4

🗳️ AI chatbots run for office

An AI-powered candidate named ‘AI Steve’ is running for U.K. Parliament in next month’s general election — creating polarizing questions around AI’s use in government affairs.

  • AI Steve is represented by businessman Steve Endacott and will appear as an independent candidate in the upcoming election.
  • Voters can interact with AI Steve online to ask policy questions and raise concerns or suggestions, which the AI will incorporate based on feedback.
  • If elected, Endacott will serve as AI Steve’s human proxy in Parliament, attending meetings and casting votes based on the AI’s constituent-driven platform.

The idea of an AI running for office might sound like a joke, but the tech behind it could actually help make our politicians more independent and (ironically) autonomous. AI-assisted governance is likely coming someday, but it’s probably still a bit too early to be taken seriously.

Source: https://www.nbcnews.com/tech/tech-news/ai-candidate-running-parliament-uk-says-ai-can-humanize-politics-rcna156991

🎤 OpenAI CTO speaks on internal models

OpenAI CTO Mira Murati participated in a Q&A with Fortune Magazine, addressing topics including criticism from Elon Musk, internal AI model strength, the 2023 board drama and more.

  • Murati said both Apple and OpenAI believe strongly in privacy and will not log data through Apple accounts or train models on user data.
  • Murati responded to Musk calling the Apple partnership ‘creepy spyware’, saying OpenAI is trying to be as transparent as possible with its approach to safety.
  • When asked about the board drama, Murati said the previous board structure ‘didn’t have accountability to anyone but themselves’.
  • The CTO also curiously said that OpenAI’s internal models ‘aren’t that far ahead’ of what the public has for free right now.

This had to be a harder-hitting Q&A than Murati expected — but her responses certainly evoked similar vibes to the infamous Sora interview. The commentary on internal models was particularly interesting, as it would imply that OpenAI doesn’t have some GPT-5 level model waiting in the wings.

Source: https://x.com/elonmusk/status/1800266437677768765

🐀 DeepMind creates a virtual rat with AI brain

Researchers from Google DeepMind and Harvard just created a virtual rodent powered by an AI neural network, capable of mimicking agile movements and neural activity of real-life rats with extreme accuracy.

  • The researchers created a bio-mechanically realistic digital rat model and trained an AI ‘brain’ to control its body in a physics simulator.
  • The AI’s brain activity patterns closely matched real rodents engaged in similar behaviors, showing it can capture principles of muscle control and movement.
  • This ‘virtual neuroscience’ approach could open new research with testing on AI animals, also transferring to engineering of advanced robotics systems.

By bridging the gap between motor control and the complexity of brains, these virtual rodents offer a new window into how neural activity generates movement. Our understanding of the brain and neural activity is about to expand exponentially with the increasing power of AI.

Source: https://www.nature.com/articles/s41586-024-07633-4

New AI Job Opportunities on June 14th 2024

  • 💼 Databricks – Sr. Product Manager: https://jobs.therundown.ai/jobs/63046708-sr-product-manager
  • 🧪 Luma AI – Research Scientist: https://jobs.therundown.ai/jobs/62973732-research-scientist
  • 📊 Dataiku – Senior Sales Engineer: https://jobs.therundown.ai/jobs/63061675-senior-sales-engineer
  • 📘 Meta – GenAI Content Manager: https://jobs.therundown.ai/jobs/60547830-genai-content-manager

What Else Is Happening in AI on June 14th 2024❗

🕵️‍♂️Former head of NSA joins OpenAI’s Safety and Security Committee

Paul M. Nakasone, a retired US Army general and a former head of the National Security Agency (NSA), will also join OpenAI’s board of directors. He will contribute to OpenAI’s efforts to better understand how AI can be used to strengthen cybersecurity by quickly detecting and responding to cybersecurity threats.

Source: https://www.theverge.com/2024/6/13/24178079/openai-board-paul-nakasone-nsa-safety

🤖Former Meta engineers launch Jace, your new autonomous AI employee

Jace uses Zeta Labs’ proprietary web-interaction model, Autonomous Web Agent-1, to use a browser to interact with websites like any human would. It allows it to handle real-world tasks like booking flights, handling hiring, or even setting up a company.

Source: https://venturebeat.com/ai/exclusive-former-meta-engineers-launch-jace-an-ai-agent-that-works-independently

💼LinkedIn is rolling out new AI-powered features for premium users

The features include searching for jobs by prompting in natural language, building a cover letter from scratch, reviewing your résumé with personalized suggestions for improving it for a specific job post, and making edits interactively with AI.

Source: https://www.theverge.com/2024/6/13/24177986/linkedin-ai-job-hunting-features-premium-subscribers

🌍Synthflow’s AI voice assistants are now multilingual!

They can fluently communicate in Spanish, German, Portuguese, French, and English. Sythflow also added corresponding voices for each language to ensure authentic and natural-sounding interactions so businesses can engage a global audience and offer personalized experiences.

Source: https://x.com/GoogleDeepMind/status/1801210428673892496

🖼️Picsart is partnering with Getty Images to develop a custom model for AI imagery

The model will be built from scratch and trained exclusively on Getty Images’ licensed creative content. It will bring responsible AI imagery to creators, marketers, and small businesses that use Picsart, enabling them to generate unique images with full commercial rights. Source: https://techcrunch.com/2024/06/13/picsart-partners-with-getty-images-to-develop-a-custom-ai-model/

A  Daily chronicle of AI Innovations June 13th 2024:

🍎 Apple to ‘pay’ OpenAI through distribution, not cash

📹 New free rival to OpenAI Sora unveiled

⌚️ Google AI gets personal with wearables

📰 Perplexity was planning revenue-sharing deals with publishers when it came under media fire

🗣️ Amazon ‘dropped the ball’ with Alexa, former employee says

🤔 Tim Cook says it’s a ‘significant possibility’ people use their iPhones less because of AI

🤖Luma AI’s dream machine debut sparks AI media wars
🖼️Stability AI’s Diffusion 3 Shakes Up Image Creation
💰No Cash Involved! Apple will pay OpenAI in distribution 

🏆NVIDIA tops MLPerf benchmarks, dominates AI training

🛠️Databricks Boosts Mosaic AI for Enterprise LLMs

💻ChromeOS gets GenAI boost via Android tech

🎬YouTube trials AI comment summaries for shorts

🤖Luma AI’s dream machine debut sparks AI media wars

LumaAI has launched Dream Machine, an AI-powered video generation system capable of generating high-quality videos via simple text prompts. The open-for-all AI tool takes about two minutes to generate a realistic five-second video clip to render specified objects and environments without compromising coherency.

Why does it matter?

The open-for-all feature of this AI tool highlights Luma AI’s approach toward democratizing generative AI and puts rival systems into the spotlight, including OpenAI’s Sora and Kuaishou’s Kling, which are only accessible to select users.

Additionally, Luma’s open approach will enable creators and companies to produce original content at a rapid speed and scale, giving it a competitive edge.

Source: https://lumalabs.ai/dream-machine

⌚️ Google AI gets personal with wearables

Google just published two new research papers, introducing an AI model that can interpret personal health data from wearable devices and an AI agent that provides detailed insights, answer health questions, and more.

  • Google’s PH-LLM is a version of Gemini fine-tuned to reason using data from wearables and generate coaching insights for sleep and fitness.
  • PH-LLM performs comparably to human experts on health insights for fitness and sleep, also achieving expert performance on certification exams.
  • Google also revealed PHIA, an AI agent that combines Gemini’s language skills with code and search capabilities to analyze wearable health data.
  • PHIA scored 84% on health insight questions, demonstrating strong reasoning and data analysis capabilities on complex queries.

Google’s research showcases AI’s potential to take health wearable tracking to the next level — going beyond generic insights to enable personalized support. Once ingrained into popular devices, these types of AI tools will put an expert fitness and nutrition coach on the wrists of every user.

Source: https://research.google/blog/advancing-personal-health-and-wellness-insights-with-ai

🖼️Stability AI’s Diffusion 3 Shakes Up Image Creation

Stability AI has launched its most sophisticated text-to-image AI open model, optimized for consumer PCs and enterprise-tier GPUs. The model delivers several noteworthy features:

  • Enables photorealistic and high-quality outputs in exceptional style by providing images with extraordinary color, lighting, and details.
  • Can comprehend long, complex prompts that involve spatial reasoning, compositional elements, actions, and styles.
  • Leverages Diffusion Transformer architecture to achieve exceptional text quality, reducing spelling, character spacing, and letter-forming errors.
  • Displays the ability to absorb minute details from small datasets, making it an ideal model for customizing and fine-tuning.

Why does it matter?

Stability’s unique features make it the most advanced text-to-image AI model. As it continues to push the boundaries of generative AI, it will be interesting to note whether it will stand out against its competitors and remain at the forefront of image generation.

Source: https://stability.ai/news/stable-diffusion-3-medium

💰No Cash Involved! Apple will pay OpenAI in distribution 

A landmark arrangement between OpenAI and Apple to integrate ChatGPT into iPhone, iPad, and Mac was unclear in its financial terms. People briefed on the matter suggest that the agreement isn’t likely to generate revenue for either party and simply provides OpenAI brand exposure to Apple’s user base while offering Apple an advanced chatbot feature.

Why does it matter?

Apple integrating ChatGPT into its devices without offering any compensation to OpenAI creates concerns among users on whether Apple intends to share their data in return.

Source: https://www.bloomberg.com/news/articles/2024-06-12/apple-to-pay-openai-for-chatgpt-through-distribution-not-cash

🍎 Apple to ‘pay’ OpenAI through distribution, not cash

  • Apple and OpenAI announced a partnership at WWDC, but according to Bloomberg, neither side is paying the other for the integration of ChatGPT into Apple’s devices.
  • The agreement between Apple and OpenAI is non-exclusive, and Apple is also negotiating with Anthropic and Google to integrate their chatbots, with a deal for Google’s Gemini expected later this year.
  • Apple plans to generate revenue from these partnerships through revenue-sharing agreements, taking a cut from AI providers that monetize their chatbot services, such as ChatGPT Plus’ $20-per-month subscription plan.
  • Source: https://www.theverge.com/2024/6/13/24177550/apple-openai-chatgpt-deal-payment-revenue-sharing-chatbot

📹 New free rival to OpenAI Sora unveiled

  • Dream Machine is a new text-to-video tool developed by Luma AI, offering a free tier you can use immediately with a Google account, though there might be waiting times due to high demand.
  • The tool creates five-second video clips at 1360×752 resolution based on user prompts, and while it has some processing delays, the outputs are impressive and close to user expectations.
  • Dream Machine offers various subscription plans, with the free option limited to 30 generations a month, and higher tiers providing more generations at different price points, up to 2,000 generations for $499.99 monthly.
  • SOurce: https://www.techradar.com/computing/artificial-intelligence/a-new-openai-sora-rival-just-landed-for-ai-videos-and-you-can-use-it-right-now-for-free

📰 Perplexity was planning revenue-sharing deals with publishers when it came under media fire

  • Perplexity, an AI search engine startup, was working on revenue-sharing deals with high-quality publishers before facing accusations from Forbes of content misuse.
  • The company aims to announce the details of these partnerships soon and has already updated its user interface to make citation of sources more prominent.
  • Perplexity’s planned revenue-sharing framework will offer media companies a recurring income alternative, contrasting with the one-off data usage deals currently offered by entities like OpenAI.
  • Source: https://www.semafor.com/article/06/12/2024/perplexity-was-planning-revenue-sharing-deals-with-publishers

🗣️ Amazon ‘dropped the ball’ with Alexa, former employee says

  • A former Amazon employee, Mihail Eric, criticized the company’s handling of Alexa, citing technical and bureaucratic problems as major hindrances to its development.
  • Eric stated that despite having ample resources and talent, Amazon mishandled its opportunity to make Alexa the leading product in conversational AI.
  • Organizational issues, such as fragmented teams and inadequate computing resources, were mentioned by Eric as significant factors that prevented Alexa from achieving its full potential.
  • Source: https://www.businessinsider.com/amazon-alexa-voice-assistant-dropped-ball-former-employee-says-2024-6

🤔 Tim Cook says it’s a ‘significant possibility’ people use their iPhones less because of AI 

  • Apple CEO Tim Cook said Apple Intelligence may reduce iPhone usage in an interview with Marques Brownlee.
  • Cook highlighted that AI will simplify previously time-consuming tasks, allowing users to complete them more efficiently and quickly.
  • Apple has added several features to the iPhone that raise awareness of how much time consumers spend on their devices and which help make their usage more intentional.
  • Source: https://www.businessinsider.com/tim-cook-people-may-use-iphones-less-because-ai-2024-6

What Else Is Happening in AI on June 13th 2024❗

🎬YouTube trials AI comment summaries for shorts: YouTube is trying out a new AI-powered comment summarizer feature on mobile apps. The feature breaks down comments into keyword-based topics, simplifying users’ engagement with Shorts with large comment sections. Source: https://support.google.com/youtube/thread/18138167

🏆NVIDIA tops MLPerf benchmarks, dominates AI training: NVIDIA has set new performance records in MLPerf Training v4.0, showcasing its dominance in generative AI and LLM training. Key achievements include tripling GPT-3 175B training performance, fine-tuning Llama 2 70B model in just 1.5 minutes using 1,024 H100 GPUs, and accelerating Stable Diffusion v2 training performance by up to 80%. Source: https://blogs.nvidia.com/blog/mlperf-training-benchmarks

🛠️Databricks Boosts Mosaic AI for Enterprise LLMs: Databricks has announced enhancements to Mosaic AI to accelerate its efforts in enterprise gen AI development. The focus lies on the development of compound AI systems, their evaluation across metrics, and governance. Releases include Mosaic AI Model Training and Agent Framework, strengthening the company’s offering against Snowflake. Source: https://venturebeat.com/data-infrastructure/databricks-bolsters-mosaic-ai-with-tools-to-build-and-evaluate-compound-ai-systems

📝No more vagueness – Adobe to update its AI terms: Adobe has promised to update its terms of service amidst user backlash, assuring that it will not train generative AI models on customer content. The company clarified that it does not own customer content and will refrain from using it to train AI models. Source: https://blog.adobe.com/en/publish/2024/06/10/updating-adobes-terms-of-use

💻ChromeOS gets GenAI boost via Android tech: Google plans to release Android-based tech stack into ChromeOS to accelerate AI innovation. The release will simplify engineering efforts, helping phones and accessories work better with Chromebooks. Through this release, Google seeks to continue rolling out its AI features to users at a faster and larger scale. Source: https://blog.chromium.org/2024/06/building-faster-smarter-chromebook.html

A  Daily chronicle of AI Innovations June 12th 2024:

🏳️ Elon Musk drops OpenAI lawsuit with no explanation

🪦 Microsoft kills off Copilot GPT Builder after just 3 months

🖼️ Stability AI launches its ‘most sophisticated’ image generator yet

🛑 Microsoft to discontinue GPT Builder from Copilot Pro

⚖️ Musk drops lawsuit alleging OpenAI strayed from its mission

☁️OpenAI leverages Oracle Cloud to extend the Microsoft Azure Al platform

🦹 Google to launch anti-theft AI features for android phones in Brazil

📱 Google expands Gemini Nano AI capabilities to Pixel 8 & 8a

💌 Yahoo Mail integrates AI for a streamlined desktop experience

🤝Samsung unites North American AI teams, appoints ex-Apple Siri executive

🏳️ Elon Musk drops OpenAI lawsuit with no explanation

  • Elon Musk has dropped his lawsuit against OpenAI and its CEO, Sam Altman, which accused the company of breaching a contract and deviating from its mission to benefit humanity.
  • The case was dismissed without prejudice, meaning Musk can refile it in the future; this decision came just a day before a scheduled hearing on OpenAI’s request to dismiss the case.
  • Musk’s lawsuit alleged a breached “Founding Agreement” with OpenAI, but the company denied this, asserting there is no binding agreement and claiming Musk sought control by merging it with Tesla.
  • Source: https://www.theverge.com/2024/6/11/24176462/elon-musk-openai-lawsuit-sam-altman-dropped

🪦 Microsoft kills off Copilot GPT Builder after just 3 months

  • Microsoft has announced the discontinuation of GPT Builder and is giving users one month to save their data before deletion.
  • GPT Builder, part of Microsoft’s Copilot Pro, allowed users to create and customize various GPTs for $20 per month but will no longer be available after July 10.
  • Despite retiring GPT Builder, Microsoft remains focused on enterprise GPT applications and will not adjust Copilot Pro subscription fees or hand the project over to the community.
  • Source: https://www.theregister.com/2024/06/11/microsoft_retires_gpt_builder/

🖼️ Stability AI launches its ‘most sophisticated’ image generator yet

  • Stability AI launched Stable Diffusion 3 Medium, described as its “most advanced text-to-image open model yet,” featuring 2 billion parameters for photorealistic output on consumer systems.
  • The new model overcomes common artefacts in hands and faces, accurately understands complex prompts, and enhances typography through Diffusion Transformer architecture.
  • Users can access SD3 Medium via Stability’s API under an open non-commercial license, while commercial usage requires contacting the startup for licensing details.
  • Source: https://thenextweb.com/news/stability-ai-launches-stable-diffusion-3-image-generator

🤖 Autonomous driving by musculoskeletal humanoids: A study

The research paper discusses the development of a musculoskeletal humanoid robot, Musashi, designed to perform autonomous driving tasks. It mimics the human body in detail with redundant sensors and a flexible body structure that are suitable for motions with complex environmental contact.

The robot is expected to sit down on the car seat, step on the acceleration and brake pedals, and operate the steering wheel by both arms.

The paper extensively discusses the hardware and learning-based software necessary for this unique approach to autonomous driving.

Why does it matter?

The learning-based software discussed in the paper highlights the adaptability of musculoskeletal humanoids. AI models could learn from both structured data and unstructured sensory inputs, potentially improving generalization and handling novel scenarios.

Source: https://arxiv.org/pdf/2406.05573

New AI Job Opportunities on June 12th 2024

  • 💼 UiPath – Strategic Sales Lead, EMEA: https://jobs.therundown.ai/jobs/62605651-strategic-sales-lead-emea
  • 💻 Captions – Frontend Engineer: https://jobs.therundown.ai/jobs/60404535-frontend-engineer-3+-years-of-experience
  • 📊 Abridge – Senior Data Analyst: https://jobs.therundown.ai/jobs/62605637-senior-data-analyst
  • 🏙️ Metropolis – Data Analyst: https://jobs.therundown.ai/jobs/62355254-data-analyst

Source: https://jobs.therundown.ai/

What Else Is Happening in AI on June 12th 2024❗

☁️ OpenAI leverages Oracle Cloud to extend the Microsoft Azure Al platform

This collaboration will provide additional capacity for OpenAI, enabling it to continue to scale. Oracle Cloud Infrastructure’s (OCI) purpose-built AI capabilities enable startups and enterprises to build and train models faster and more reliably anywhere in Oracle’s distributed cloud.Source: https://www.oracle.com/news/announcement/openai-selects-oracle-cloud-infrastructure-to-extend-microsoft-azure-ai-platform-2024-06-11 

🦹 Google to launch anti-theft AI features for android phones in Brazil

The features include three lock modes: AI Movement Detection (recognizes common theft-related movements), Remote Lock, and Internet Access Timeout (automatically locks the screen if the device lacks internet access for an extended period). Starting in July 2024, Brazilian users with Android phones running version 10 or higher will have access to these features. Source: https://www.reuters.com/technology/artificial-intelligence/google-test-anti-theft-ai-feature-phones-brazil-2024-06-11

📱 Google expands Gemini Nano AI capabilities to Pixel 8 & 8a

This update includes features like “Summarize in Recorder” and “Smart Reply.” Users need to access developer options to enable these features. Additionally, the update enhances Find My Device functionality and allows Display Port output via USB-C. Older Pixel devices, such as the Pixel Fold and Pixel 6, also receive new features. https://www.theverge.com/2024/6/11/24176127/google-pixel-feature-drop-june-gemini-nano

💌 Yahoo Mail integrates AI for a streamlined desktop experience

The enhancements include AI-generated email summaries, a “Priority” inbox tab for urgent messages, quick action buttons, and the ability to link other email accounts within Yahoo Mail. Existing users can opt in, while new users have immediate access. Source: https://www.theverge.com/2024/6/11/24175280/yahoo-mail-generative-ai-gmail

🤝Samsung unites North American AI teams, appoints ex-Apple Siri executive

Samsung merged its AI research centers in Toronto and California into a single unit, the North America AI Center. To lead this new North American AI Center, Samsung recruited Murat Akbacak, a former Siri executive at Apple. Source: https://www.bloomberg.com/news/articles/2024-06-11/samsung-taps-apple-siri-veteran-to-lead-north-american-ai-group

A  Daily chronicle of AI Innovations June 11th 2024:

🍎 Apple brings ChatGPT to Siri

🖥️ Apple reveals macOS Sequoia with iPhone mirroring, new Passwords app, and more

⚡ Elon Musk threatens Apple device ban after OpenAI partnership

🍎 Apple introduces Apple Intelligence, its new Gen AI tech
🚫 Musk threatens to ban Apple devices over ChatGPT integrations
💼 OpenAI hires Nextdoor and Instagram veterans as CFO and CPO

🍎 Apple brings ChatGPT to Siri

  • Apple introduced AI updates for Siri at WWDC 2024, enabling it to perform actions across apps, manage notifications, automatically write and summarize text, and understand user context with “onscreen awareness.”
  • Siri’s new capabilities include referencing one app to perform actions in another, such as playing a podcast your partner sent, changing the tone of written text with a prompt, and allowing users to interact by typing as well as speaking.
  • Apple emphasizes privacy with on-device processing for many features, while more complex requests are handled by “Private Cloud Compute” on Apple Silicon-powered servers, with independent experts verifying the code running on its servers.
  • Source: https://www.theguardian.com/technology/article/2024/jun/10/apple-ai-product-launch

🖥️ Apple reveals macOS Sequoia with iPhone mirroring, new Passwords app, and more

  • Apple announced macOS Sequoia, featuring major updates such as iPhone mirroring that allows users to control their iPhone from their Mac.
  • Sequoia introduces a stand-alone Passwords app, syncable with other Apple devices and Windows PCs via iCloud, along with new Safari improvements like Highlights for extracting useful information.
  • Apple also revealed Game Porting Toolkit 2 for better Windows game compatibility, and the macOS Sequoia beta will be available this month with a full release expected in the fall.
  • Source: https://techcrunch.com/2024/06/10/apple-unveils-macos-sequoia

⚡ Elon Musk threatens Apple device ban after OpenAI partnership

  • Elon Musk has threatened to ban Apple devices at his companies because Apple plans to integrate OpenAI’s ChatGPT into its products.
  • Musk argues that allowing OpenAI access through Apple devices poses a security risk and criticizes Apple for not developing its own AI technology.
  • Musk’s discontent also stems from his ongoing disputes with OpenAI, including past legal actions and competitive tensions with his own AI ventures like xAI and its Grok chatbot.
  • Source: https://www.theregister.com/2024/06/11/musk_wants_to_ban_apple/

Apple introduces Apple Intelligence, its new Gen AI tech

Apple unveiled Apple Intelligence at WWDC 2024, a personal intelligence system deeply integrated into iOS 18, iPadOS 18, and macOS Sequoia. Leveraging Apple silicon’s power, it understands and creates language/images, takes actions across apps, and simplifies tasks using personal context. With Private Cloud Compute, Apple sets a new privacy standard in AI by flexibly processing between on-device and dedicated Apple silicon servers.

Apple Intelligence will power new features like AI-generated Bitmojis, natural language images, and video search, as well as enhanced integration with apps and Siri.

Apple also partnered with OpenAI to integrate ChatGPT with its apps, including Siri. This will allow users to directly access GPT-4 through Apple’s platforms. Additionally, Apple confirmed plans to work with other models in the future, including Google’s Gemini. The iPhone 15 Pro and devices with M1 or newer chips will be the first to receive the full suite of Apple Intelligence features.

Siri Upgrades
  • A next-gen Siri will converse more naturally, remember context across requests, and accomplish more complex tasks by better understanding both voice and text.
  • Siri also gains ‘onscreen awareness’, with the ability to take actions and utilize on-device info to better tailor requests to the individual user.
New AI Features
  • New AI writing tools built into apps like Mail, Messages, and Notes will allow users to auto-generate and edit text.
  • Mail will utilize AI to better organize and surface content in inboxes, while Notes and Phone gain new audio transcription and summarization capabilities.
  • AI-crafted ‘Genmojis’ enable personalized text-to-image emojis, and a new “Image Playground” feature introduces an image generation tool from prompts.
  • Photos get more conversational search abilities, the ability to create photo ‘stories’, and new editing tools.
Privacy
  • A focus of the AI reveal was privacy — with new features leveraging on-device processing when possible and Private Cloud Compute for more complex tasks.
  • Private Cloud Compute (PCC) is Apple’s new intelligence system specifically for private AI processing on the cloud.
  • The new AI features will be opt-in, so users will not be forced to adopt them.
OpenAI Integration
  • The OpenAI partnership will allow Siri to leverage ChatGPT/GPT-4o when needed for more complex questions.
  • OpenAI’s blog also outlined additional ChatGPT tools like image generation and document understanding embedded into the new OS.

Why does it matter?

This is generative AI as only Apple can deliver it– put at the core of its products, accessing users’ personal data for truly helpful AI assistance but keeping it secure through on-device processing and Private Cloud Compute. Apple’s long-known focus on user privacy + exceptional UX could inspire a new era of AI development.

Source: https://techcrunch.com/2024/06/11/everything-apple-announced-wwdc-2024/

Musk threatens to ban Apple devices over ChatGPT integrations 

Elon Musk threatens to ban iPhones and other Apple devices from his companies, such as Tesla, SpaceX, and xAI, over Apple’s announcement of ChatGPT integration into its OS. Musk thinks this integration will allow ChatGPT to access too much private user data through Apple devices. He claims Apple doesn’t understand OpenAI’s tech and is putting user privacy at risk.

However, Apple and OpenAI stated that users will be asked permission before any data is sent to ChatGPT.

Why does it matter?

This highlights Musk’s skepticism about the privacy implications of Apple’s AI strategy. But it could significantly impact Apple users in the future, as they may no longer be able to use iPhones and other Apple devices at Musk’s companies

Source: https://techcrunch.com/2024/06/10/elon-musk-threatens-to-ban-apple-devices-from-his-companies-over-apples-chatgpt-integrations

OpenAI hires Nextdoor and Instagram veterans as CFO and CPO

OpenAI has hired two experienced executives to fill important leadership roles. Sarah Friar, previously the CEO of the neighborhood app Nextdoor, was appointed OpenAI’s new chief financial officer (CFO). Kevin Weil, a former Instagram and Twitter executive who oversees product teams, has been named OpenAI’s chief product officer (CPO).

As CFO, Friar will manage OpenAI’s financial operations and investments as the company expands globally. As CPO, Weil will lead product development utilizing OpenAI’s AI research to create new products for consumers and enterprise customers.

Why does it matter?

This indicates OpenAI’s ambition to scale as it commercializes its AI technology. It also suggests that OpenAI is solidifying its position as a major force in the AI industry while navigating the recently surfaced challenges around AI safety.

Source: https://openai.com/index/openai-welcomes-cfo-cpo

🐶 Researchers teach AI to speak dog

A new study from the University of Michigan leveraged AI models trained on human speech to decode the meaning behind dog noises — identifying details like breed, age, gender, and emotional state with 70% accuracy.

  • Researchers gathered vocalizations from 74 dogs of varying breeds, ages, and situational contexts.
  • The noises were inputted into an AI model originally designed to analyze human voices, trained on 960 hours of speech, and fine-tuned for dogs.
  • The AI was able to predict individual dogs from barks, distinguish dog breed and gender, and match barks to emotional context like play and aggression with 70% accuracy.

AI is not only bridging the language gap for humans across the globe — but also potentially across species as well. Communicating with other intelligent animals (or at least better understanding them) seems like a skill issue that’s going to be solved sooner rather than later.

Source: https://arxiv.org/pdf/2404.18739?

New AI Job Opportunities on June 11th 2024

  • 🔧 Fiddler AI – Technical Project Manager, Customer Success: https://jobs.therundown.ai/jobs/62351203-technical-project-manager-customer-success-remote-east-coast
  • 🎓 Cohere – Research Intern (PhD): https://jobs.therundown.ai/jobs/62351142-research-intern-phd
  • 🔬 Meta – Research Engineer, Conversational AI (Reality Labs): https://jobs.therundown.ai/jobs/62419977-research-engineer-conversational-ai-reality-labs
  • 🎨 Palantir Technologies – Product Designer, US Government: https://jobs.therundown.ai/jobs/62398950-product-designer-us-government

What Else Is Happening in AI on June 11th 2024❗

⏳ AI could save financial advisors 10-15 hours per week

Morgan Stanley is using an AI tool to transcribe and categorize client conversations, saving financial advisors 10-15 hours per week. It allows advisors to better prepare for client meetings by reminding clients of past discussions and anticipating their needs. The AI tool can automatically summarize meetings and generate follow-up emails.

Source: https://www.reuters.com/technology/morgan-stanley-ceo-says-ai-could-save-financial-advisers-10-15-hours-week-2024-06-10

🤖 Anthropic reveals Claude 3’s character training process

Anthropic has developed a “character training” process for their AI assistant, Claude, to imbibe traits like curiosity, open-mindedness, and thoughtfulness. It aims to make Claude more discerning in their interactions and able to express their own views while also displaying openness to other perspectives.

Source: https://www.anthropic.com/research/claude-character

👑 FanVue introduces the world’s first AI beauty pageant contest 

Influencer marketing platform FanVue hosted the world’s first AI-powered beauty pageant, “Miss AI,”. Over 1,500 AI-generated digital models worldwide compete for the title and a $20,000 prize pool. The contest aims to showcase the capabilities of AI in creating photorealistic digital avatars that can function as influencers and brand ambassadors.

Source: https://www.npr.org/2024/06/09/nx-s1-4993998/the-miss-ai-beauty-pageant-ushers-in-a-new-type-of-influencer

🔍 Perplexica provides an open-source AI search alternative for Perplexity

Perplexica uses LLMs, machine learning algorithms, and embedding models to provide refined search results and clear answers from cited sources. Perplexica offers multiple modes, like various “Focus Modes” tailored for specific question types.

Source: https://github.com/ItzCrazyKns/Perplexica

🧒 AI tools secretly use real images of children for training 

Over 170 images and personal details of Brazilian children have been repurposed by an open-source AI training dataset called LAION-5B without the children’s consent. The images were originally posted on mommy blogs and personal social media. Researchers are concerned these images could be used to generate deepfakes or reveal sensitive information.

Source: https://www.wired.com/story/ai-tools-are-secretly-training-on-real-childrens-faces

How does Apple send your data to its cloud AI servers? Very carefully, it claims. https://www.engadget.com/how-does-apple-send-your-data-to-its-cloud-ai-servers-very-carefully-it-claims-233312425.html

Apple confirms plans to work with Google’s Gemini ‘in the future’. https://techcrunch.com/2024/06/10/apple-confirms-plans-to-work-with-googles-gemini-in-the-future/

Paris-based AI startup Mistral AI raises $640 million. https://techcrunch.com/2024/06/11/paris-based-ai-startup-mistral-ai-raises-640-million/

Tim Cook is ‘not 100 percent’ sure Apple can stop AI hallucinations: https://www.theverge.com/2024/6/11/24176035/tim-cook-apple-stop-ai-hallucinations

Nvidia shipped 3.76M data center GPUs in 2023 — dominates business with 98% revenue share. https://www.tomshardware.com/tech-industry/nvidia-shipped-376m-data-center-gpus-in-2023-dominates-business-with-98-revenue-share

A  Daily chronicle of AI Innovations June 10th 2024:

⚖️ Meta faces legal complaints over AI data plans

🚀 Alibaba’s Qwen2 AI models outperform GPT-4 & Llama-3
🧠 SAP & Nvidia are developing applications with AI & digital twins
🕵️‍♂️ Chinese tech giants exploit Nvidia AI chip loophole

🍎 Apple launches “Apple Intelligence” at WWDC 2024 for iPhone, iPad, and Mac

🚀 TCS launches TCS AI WisdomNext™, an industry-first GenAI aggregation platform

🤝 Human Native AI is building a marketplace for AI training licensing deals

🤖 Hugging Face and Pollen Robotics launched an open-source robot for household chores

Alibaba’s Qwen2 AI models outperform GPT-4 & Llama-3

Alibaba launched Qwen2 with five sizes ranging from 0.5B to 72B parameters. These models are trained in 27 additional languages besides English and Chinese, showcasing state-of-the-art performance in benchmarks. The models deliver significantly improved performance in coding and mathematics and extended context length support up to 128K tokens. Despite having fewer parameters, qwen2-72 B outperforms leading models like Llama-3-70B and its predecessor Qwen1.5-110B.

Qwen2-72B-Instruct performs comparably to GPT-4 in terms of safety and significantly outperforms Mistral-8x22B. The models are released under Apache 2.0 and Qianwen License on Hugging Face and ModelScope.

Why does it matter?

Qwen2 beats Meta’s model despite being trained on relatively fewer tokens. The researchers attribute it to more efforts put into data cleaning and training, implying innovative approaches on their end.

However, it also signals the slow shift in how LLMs are developed– from solely relying on quantity of data to prioritizing the quality of data and training techniques.

Source: https://qwenlm.github.io/blog/qwen2

SAP & Nvidia are developing applications with AI & digital twins

At SAP’s Sapphire event in Orlando, Florida, SAP and NVIDIA announced their collaboration to enhance SAP’s generative AI copilot, Joule, with two new capabilities: SAP Consulting and ABAP Developer. These new features are powered by NVIDIA AI Enterprise software.

Additionally, SAP is integrating NVIDIA Omniverse Cloud APIs into its Intelligent Product Recommendation solution to simplify the buying and selling process for complex products. This integration will allow salespeople to visualize 3D product digital twins directly within the SAP Intelligent Product Recommendation interface, making it easier to understand the products.

Why does it matter?

Using NVIDIA Omniverse Cloud APIs in SAP’s Intelligent Product Recommendation solution accelerates the quote generation process and increases sales and customer satisfaction by enabling sales representatives to provide more accurate, tailored recommendations.

Source: https://blogs.nvidia.com/blog/sap-sapphire-ai-omniverse

Chinese tech giants exploit Nvidia AI chip loophole

The U.S. government prohibits Nvidia from selling A.I. chips directly to Chinese companies due to national security concerns. Still, ByteDance is accessing Nvidia’s A.I. chips for its U.S. operations by leasing them from Oracle, as the current U.S. rules do not explicitly prohibit Chinese companies from accessing the chips if used within the U.S.

Other Chinese tech giants like Alibaba, Tencent, and China Telecom seek similar arrangements with U.S. cloud providers. The U.S. Commerce Department proposed a rule to tighten controls, but it faced opposition from cloud providers and remains in limbo.

Why does it matter?

Even if the loophole is closed, Alibaba and Tencent have discussed obtaining Nvidia chips for their U.S.-based data centers. It could further escalate the AI “arms race” and rivalry between the USA and China as both nations seek to outpace each other in developing advanced AI systems for economic and military advantages.

Source: https://www.theinformation.com/articles/chinas-nvidia-loophole-how-bytedance-got-the-best-ai-chips-despite-u-s-restrictions

What Else Is Happening in AI on June 10th 2024❗

🍎 Apple will launch “Apple Intelligence” at WWDC 2024 for iPhone, iPad, and Mac

Leaks suggest Apple will reveal “Apple Intelligence”, aka AI, at the WWDC event this week. These AI features will focus on broad appeal and privacy, with opt-in not mandatory. Apple will use its own tech and OpenAI tools to power the new AI features.

Source: https://www.theverge.com/2024/6/7/24173528/apple-intelligence-ai-features-openai-chatbot

🚀 TCS launches TCS AI WisdomNext™, an industry-first GenAI aggregation platform

The platform allows organizations to compare and experiment with GenAI models across cloud services in a single interface. It offers ready-to-deploy business solution blueprints with built-in guardrails for quick adoption.

Source: https://www.tcs.com/who-we-are/newsroom/press-release/tcs-launches-wisdomnext-an-industry-first-genai-aggregation-platform

🚨 A study by Harvard, MIT, and Wharton reveals junior staff is not reliable for AI training

Junior consultants who participated in a GPT-4 experiment struggled with AI risk mitigation, with their tactics lacking a deep understanding of the technology and focusing on changing human behavior rather than AI system design. The findings highlight the need for top-down AI governance, expert input, and upskilling across all levels of the organization.

Source: https://venturebeat.com/ai/harvard-mit-and-wharton-research-reveals-pitfalls-of-relying-on-junior-staff-for-ai-training

🤝 Human Native AI is building a marketplace for AI training licensing deals

The platform helps AI companies find data to train their models while ensuring rights holders are compensated. Rights holders upload content for free and connect with AI companies for revenue share or subscription deals. Human Native AI helps prepare and price content, monitors for copyright infringements, and takes a cut of each deal.

Source: https://techcrunch.com/2024/06/08/deal-dive-human-native-ai-is-building-the-marketplace-for-ai-training-licensing-deals

🤖 Hugging Face and Pollen Robotics launched an open-source robot for household chores

The humanoid Reachy2 was initially controlled by a human wearing a VR headset. Then, a machine learning algorithm studied the teleoperation sessions to learn how to perform the tasks independently. The dataset and trained model used for the demo are open-sourced on Hugging Face, allowing anyone to replicate the process on smaller robots at home.

Source: https://venturebeat.com/ai/hugging-face-and-pollen-robotics-show-off-first-project-an-open-source-robot-that-does-chores

👑 The world’s first AI beauty pageant

The World AI Creator Awards and creator platform FanVue are currently running the inaugural ‘Miss AI’ contest, with over 1,500 AI-generated models competing for the world’s first AI beauty pageant.

  • 10 finalists were selected from a pool of 1,500 AI-generated contestants, with the winner set to be announced at the end of June.
  • The AI models ‘hail’ from countries across the globe, showcasing various causes and personalities in addition to their photorealistic images.
  • Judges will also evaluate the AI tech behind the avatars, including prompts/image outputs and the creator’s ability to engage audiences on social media.
  • The prize pool includes $20,000 in awards and access to PR and mentorship programs.

With how good AI image generation capabilities have become, we’re likely already being exposed to an influx of generated brand ambassadors and models without even realizing it. While the idea of an AI beauty pageant may sound strange… it’s probably only going to get weirder from here.

Source: https://www.npr.org/2024/06/09/nx-s1-4993998/the-miss-ai-beauty-pageant-ushers-in-a-new-type-of-influencer

🧠 Microsoft Recall gets safety changes

Microsoft is making changes to its new Recall on-device AI feature following security concerns — with the tool now starting off by default and gaining new encryption protections ahead of the release on June 18.

  • The Copilot + Recall feature was unveiled at the Microsoft Build event in May, which will run locally and take constant screenshots to remember user actions.
  • The company faced backlash after experts warned of the tool’s ‘privacy nightmare’ with the potential to expose screenshots and activity to hackers.
  • The feature will now be off by default, with users having to opt in manually — with new encryption and authentication being added for extra safety measures.
  • The tool will roll out in ‘preview’ on the new Copilot + PCs set to launch on June 18.

While the Recall feature was one of the flashiest features revealed at Build — it also is one of the most controversial. While the screenshots enable memory and get us a step closer to the AI agent dream, its also a tough hurdle to get past for increasingly privacy-focused consumers.

Source: https://www.therundown.ai/p/microsofts-total-recall

🔊 AI TRAINING: Using ‘Background Conversations’ in ChatGPT

ChatGPT’s latest update introduced the “Background Conversations” feature, allowing you to continue your chat even when using other apps or when your screen is off.

  1. Install the latest ChatGPT app update on your phone.
  2. Open the app and go to Settings.
  3. Scroll to “Voice Mode” and toggle “Background Conversations” on.
  4. Use ChatGPT hands-free while multitasking or with your screen off.

Note: If you still don’t see the option, the feature may not have rolled out to your account yet.

🧠 AI RESEARCH: Concise prompting cuts AI costs by 20%

Researchers just found that adding a simple ‘be concise’ instruction to chain-of-thought prompts can reduce the length of AI responses by 50% with minimal impact on accuracy — leading to over 20% cost savings on API calls.

  • Chain-of-thought (CoT) prompting has the AI explain its reasoning step-by-step before giving a final answer, improving accuracy on complex problems.
  • Researchers tested adding ‘be concise’ to CoT prompts on GPT 3.5 and 4, using 1,000 multiple-choice questions across 10 topics.
  • The instruction made responses about 50% shorter on average vs. normal CoT prompting, with no significant accuracy impact.
  • The approach also reduced per-query API costs by over 20% for both models, with the more concise prompts generating fewer tokens.

Next time your AI chatbot is getting a little too wordy, this one simple prompt trick is all you need! Despite the massive complexities and uncertainties surrounding how LLMs work, sometimes these strange, seemingly basic quirks make all the difference in getting the optimal outputs.

Source: https://arxiv.org/pdf/2401.05618

New AI Job Opportunities on June 10th 2024:

  • 🔬 Meta – Research Engineer, Conversational AI (Reality Labs): https://jobs.therundown.ai/jobs/62419977-research-engineer-conversational-ai-reality-labs
  • 🎨 Palantir Technologies – Product Designer, US Government: https://jobs.therundown.ai/jobs/62398950-product-designer-us-government
  • 📊 Databricks – Strategic Sourcing Manager, Technology: https://jobs.therundown.ai/jobs/62399340-strategic-sourcing-manager-technology?
  • 📈 Anyscale – Marketing Operations Manager: https://jobs.therundown.ai/jobs/62367786-marketing-operations-manager

A  Daily chronicle of AI Innovations June 07th 2024:

🎧 OpenAI reverse engineers the workings of AI models
🤖 New Chinese video generation model beats OpenAI’s Sora
📊 Nvidia is now the second-most valuable company, overtaking Apple

🚀Adobe’s launches AEP AI Assistant to help brands master customer data and outreach

🏆The most comprehensive benchmarking & leaderboard for image models is here!

🧬AI used to predict potential new antibiotics in groundbreaking study

🔒Meta gets EU complaints from 11 countries over use of personal data to train AI

🔍The Federal Trade Commission is investigating Microsoft-Inflection AI deal

OpenAI reverse engineers the workings of AI models

In new research, OpenAI has shared improved methods for finding a large number of “features”—patterns of activity in AI models that are human interpretable. They developed new state-of-the-art methodologies that allow scaling sparse autoencoders to tens of millions of features on frontier AI models.

It demonstrated smooth and predictable scaling, with better returns to scale than prior techniques. And they could find 16 million features in GPT-4. The research also introduces several new metrics for evaluating feature quality.

OpenAI has shared the paper, code, and feature visualizations to foster further exploration.

Why does it matter?

It could bring us closer to understanding the inner workings of neural networks. This interpretability will eventually increase model trustworthiness and steerability.

While OpenAI is being criticized for taking unnecessary risks with AI technology, this shows it is also keen on tackling the risk by making models more explainable.

Source: https://openai.com/index/extracting-concepts-from-gpt-4

New Chinese video generation model beats OpenAI’s Sora

Kuaishou, a Chinese tech company, has introduced Kling, an AI model for video generation. It can make videos up to two minutes long at 1080p resolution and 30 frames per second, vs. Sora’s one-minute videos.

Kuaishou claims Kling correctly simulates the physical properties of the real world, including complex motion sequences. Using a diffusion transformer, it can also combine concepts and create fictional scenes, such as a cat driving a car through a busy city.

The model is currently available as a public demo in China.

Why does it matter?

Move over, Sora. This Chinese alternative is blowing everyone’s minds. It is becoming increasingly clear that China is catching up fast with the USA in AI.

Source: https://kling.kuaishou.com/

Nvidia is now the second-most valuable company, overtaking Apple

Nvidia rallied to record highs on Wednesday, with it’s stock market valuation hitting $3 trillion and overtaking Apple to become the world’s second most valuable company. This comes after Nvidia made a series of major announcements in the past week.

However, Nvidia’s stock has surged 147% so far in 2024, with demand for its top-of-the-line processors far outstripping supply as Big Tech races to build out their AI computing capabilities and dominate the emerging technology.

Microsoft remains the world’s most valuable company, with a market value of approximately $3.15 trillion.

Why does it matter?

Nvidia is now seen as a key company to watch to see how fast AI-powered tech is spreading across the business world, a shift that Nvidia’s founder, Jensen Huang, has declared as the dawn of the “next industrial revolution”.

Source: https://www.bbc.com/news/articles/c6ppqd3j621o

🎥 Kuaishou launches new Sora rival
Chinese tech firm Kuaishou just introduced KLING, a new text-to-video AI model capable of generating high-quality videos up to 2 minutes long with outputs that appear to rival OpenAI’s still-unreleased Sora.
The details:
KLING can produce videos at 1080p resolution with a maximum length of 2 minutes, surpassing the 1-minute Sora videos demoed by OpenAI.
KLING’s demos include realistic outputs like a man eating noodles and scenic shots, as well as surreal clips like animals in clothes.
The model uses a 3D space-time attention system to simulate complex motion and physical interactions that better mimic the real world.
The model is currently available to Chinese-based users as a public demo on the KWAI iOS app.
Why it matters: These generations are even more mind-blowing when you consider that Will Smith’s spaghetti-eating abomination was barely a year ago. With users still anxiously waiting for the public release of Sora, other competitors are stepping in — and the AI video landscape looks like it’s about to heat up in a major way.
Source:

🇨🇳 ByteDance’s AI chip loophole

TikTok parent company ByteDance is renting advanced Nvidia AI chips and using them on U.S. soil, exploiting a loophole to sidestep restrictions on China’s AI chip exports.
The details:
Due to national security concerns, the U.S. government prohibits Nvidia from selling AI chips like the A100 and H100 directly to Chinese companies.
The restrictions don’t prevent Chinese firms from renting chips for use within the U.S. — ByteDance is allegedly leasing servers with chips from Oracle.
ByteDance reportedly had access to over 1,500 H100 chips and several thousand A100s last month through the Oracle deal.
Other Chinese giants like Alibaba and Tencent are also reportedly exploring similar options, either renting from U.S. providers or setting up US data centers.
Why it matters: The AI race between the U.S. and China is only escalating — and it appears major players are going to get AI chips by any means necessary. While the U.S. tries to stall its rival’s progress with restrictions, it feels like a game of whack-a-mole that won’t stop China from reaching its AI goals.
Source:

🔍 AI RESEARCH: OpenAI probes GPT’s inner workings

OpenAI just released a new paper detailing a method for reverse engineering concepts learned by AI models and better understanding ChatGPT’s inner workings.
The details:
The paper was authored by members of the recently disbanded superalignment team, including Ilya Sutskever and Jan Leike.
‘Scaling and Evaluating Sparse Autoencoders’ outlines a technique to ID patterns representing specific concepts inside GPT-4.
By using an additional model to probe the larger model, researchers found a way to extract millions of activity patterns for further exploration.
OpenAI released open-source code and a visualization tool, allowing others to explore how different words and phrases activate concepts within models.
Why it matters: Much like Anthropic’s recent “Golden Gate Claude” and corresponding research, AI firms are still working to understand what’s truly going on underneath the hood. Cracking AI’s black box would be a big step towards better safety, tuning, and controllability of rapidly advancing models.
Source:

New AI Job Opportunities on June 07th 2024:

📊 Harvey – Data Analyst
🧠 Meta – Applied AI Research Scientist – Reinforcement Learning
📣 DeepL – Corporate Communications Lead
🤝 Palantir Technologies – Enablement Partner

What Else Is Happening in AI on June 07th 2024❗

🚀Adobe’s launches AEP AI Assistant to help brands master customer data and outreach

Adobe’s Adobe Experience Platform (AEP) is a real-time customer experience management (CXM) solution designed for enterprises. AEP AI Assistant is a gen AI chatbot hooked up to the AEP and an enterprise’s storehouse of advertising and customer data, brand assets, and content collateral (at their direction). 

Source: https://venturebeat.com/ai/adobes-new-aep-ai-assistant-is-here-to-help-brands-master-customer-data-and-outreach/

🏆The most comprehensive benchmarking & leaderboard for image models is here!

AI image models have achieved near-photographic quality. But how do they compare? Are the open-source alternatives on par with their proprietary counterparts? The Artificial Analysis Text to Image Leaderboard aims to answer these questions with human preference-based rankings.

Source: https://huggingface.co/blog/leaderboard-artificial-analysis2

🧬AI used to predict potential new antibiotics in groundbreaking study

Scientists used an algorithm to mine the entirety of the microbial diversity that we have on earth – or a huge representation of that – and find almost 1million new molecules encoded or hidden within all that microbial dark matter.

Source: https://www.theguardian.com/society/article/2024/jun/05/ai-antibiotic-resistance

🔒Meta gets EU complaints from 11 countries over use of personal data to train AI

Meta rejected the criticism and referred to a May 22 blog in which it said it uses publicly available online and licensed information to train AI as well as information that people have shared publicly on its products and services.

Source: https://www.reuters.com/technology/meta-gets-11-eu-complaints-over-use-personal-data-train-ai-models-2024-06-06

🔍The Federal Trade Commission is investigating Microsoft-Inflection AI deal

The FTC has sent subpoenas to tech giant and startup, asking whether their partnership evaded the required government antitrust review of the transaction. Microsoft picked off Inflection’s specialized workforce of AI researchers but didn’t purchase the company outright. 

Source: https://www.wsj.com/tech/ai/ftc-opens-antitrust-probe-of-microsoft-ai-deal-29b5169a

A  Daily chronicle of AI Innovations June 06th 2024:

🎧 Stability AI’s sound generator creates drum beats and instrument riffs
🤖 xAI to build the gigafactory of compute
📊 New study reveals key findings on young peoples’ use of Gen AI

💊 AI predicts nearly 1M new antibiotics

🔊 Stability AI releases open audio model

🛠️Mistral’s new SDK simplifies customization for users 

🏢Salesforce inaugurates its first AI Centre in London

🛒eBay unveils new AI-tool to enhance product image backdrops

👥Asana introduces AI teammates

🚀Writer launches a suite of No-Code development tools 

🎧 Stability AI’s sound generator creates drum beats and instrument riffs

Stability AI’s Stable Audio Open can generate up to 47-second audio samples based on text descriptions. The open AI model is trained on data from 486,000 samples of royalty-free music samples. The tool enables users to generate drum beats, instrument riffs, and ambient sounds.

However, the AI model has its limitations.

  • It is unable to produce full songs, melodies, or vocals.
  • Its terms of service prohibit users from using Stable Audio Open commercially.
  • Its training data is biased toward the English language and specific music styles.

Why does it matter?

The tool will enable creative users like sound designers, musicians, and developers to fine-tune the model to their own custom audio data, which is a major leap in the realm of generative AI’s audio generation capabilities.

Source: https://stability.ai/news/introducing-stable-audio-open

🤖 xAI to build the gigafactory of compute

The AI startup seeks to build the world’s largest supercomputer in a multi-billion dollar project. The company plans to use this supercomputer to develop AI products, including its chatbot, Grok. The facility will be powered by Nvidia’s H100 GPUs. The project aims to be operational by fall 2025.

The project will reportedly use Nvidia AI chips originally intended for Tesla, raising concerns about conflicts of interest. Moreover, Musk hasn’t yet delivered Grok 2, an advanced AI model that he had promised in May.

Why does it matter?

The supercomputer’s support for the development of Grok could become a groundbreaking advancement in the field of AI technology. Moreover, the  project represents a multi-billion dollar investment that could create numerous job opportunities and boost the economy.

Source: https://memphischamber.com/velocity-meets-potency-xai-announces-memphis-as-new-home

📊 New study reveals key findings on young peoples’ use of Gen AI

The study directly involved young readers and examined the use of generative AI by use, ethnicity, age, gender, and LGBTQ+ identity. Key findings include:

  • 50% of the survey respondents (aged 14-22) have used generative AI. However, only 4% use it daily.
  • For 53% of respondents, the use case for generative AI was obtaining information, while for 51%, it was brainstorming.
  • Black young people are more likely to use generative AI compared to their white peers. Reasons include getting information, brainstorming ideas, and assistance with schoolwork.
  • Young people of Latin origin are more likely than white people to use generative AI for multiple purposes, including image generation and getting help with their jobs.
  • Out of respondents who have never used generative AI, 34% believed it would not be helpful.
  • Among people never having used generative AI, LGBTQ+ young people are more likely to use it in comparison to cisgender and straight respondents.
  • 41% of respondents believe that generative AI will have a positive as well as negative impact on their lives in the next 10 years.

Why does it matter?

Understanding the diverse ways in which the younger users engages with generative AI enables businesses to uncover trends and data-oriented insights to improve their AI models and deliver, user experiences that are responsive and inclusive.

Source: https://www.commonsensemedia.org/sites/default/files/research/report/teen-and-young-adult-perspectives-on-generative-ai.pdf?

💊 AI predicts nearly 1M new antibiotics

Researchers just published a new study detailing the use of AI to predict close to 1M new antibiotics hidden within tiny microbes all over the world, uncovering new potential treatments against bacteria and superbugs.

  • Researchers used AI to analyze publicly available data on over 100,000 different genomes and meta-genomes.
  • The AI then predicted which parts of the microbial genomes could potentially produce antibiotic compounds, generating a list of nearly one million candidates.
  • 100 of the AI-predicted drug candidates were tested in the lab, with 79 of them being a potential antibiotic.
  • The paper’s author Cesar de la Fuente said the findings are “the largest antibiotic discovery ever”, accelerating the process from years to just hours.

As the world faces growing threats from antibiotic-resistant bacteria, AI’s ability to unlock millions of new potential treatments could be a lifeline toward staying ahead in the race to outsmart superbugs responsible for millions of deaths every year.

Source: https://www.theguardian.com/society/article/2024/jun/05/ai-antibiotic-resistance

🔊 Stability AI releases open audio model

Stability AI just unveiled Stable Audio Open, a new model enabling users to create short audio samples and sound effects from text prompts.

  • Stable Audio Open can create up to 47 seconds of high-quality audio from text prompts, focused primarily on sound effects and production elements.
  • The model was trained on data from FreeSound and Free Music Archives to avoid copyright issues.
  • Stable Audio Open can also be fine-tuned on custom audio, allowing artists to train the model on their own samples.

While the AI advances in text-to-image models have been the most visible (literally), both video and audio are about to take the same leap. Putting these tools in the hands of creatives will redefine traditional workflows — from musicians brainstorming new beats to directors crafting sound effects for film and TV.

Source: https://stability.ai/news/introducing-stable-audio-open

New AI Job Opportunities on June 06th 2024

  • 💼 Twelve Labs – Head of Finance: https://jobs.therundown.ai/jobs/61919747-head-of-finance
  • 🔒 Fiddler AI – Senior Security Engineer: https://jobs.therundown.ai/jobs/61902058-senior-security-engineer-remote-or-hybrid
  • 🎯 Siena – Director of Customer Success: https://jobs.therundown.ai/jobs/61919768-director-of-customer-success
  • 💻 Cresta – Senior Software Engineer (AI Agent): https://jobs.therundown.ai/jobs/61882269-senior-software-engineer-ai-agent

What Else Is Happening in AI on June 06th 2024❗

🛠️Mistral’s new SDK simplifies customization for users 

Mistral’s SDK, Mistral-Finetune, will allow users to modify its models on workstations, servers, and small data center nodes. While optimized for multi-GPU setups, the SDK can also scale down to a single Nvidia A100 or H100 GPU. It enables developers to fine-tune data sets such as UltraChat within half an hour, says Mistral. (https://techcrunch.com/2024/06/05/mistral-launches-new-services-sdk-to-let-customers-fine-tune-its-models)

🏢Salesforce inaugurates its first AI Centre in London

The goal of the centre is to foster collaboration, AI training, upskilling, and create AI-related jobs among developers, industry experts, and customers. This step is a part of its £4bn five-year investment in the UK, solidifying London’s position as a global AI hub. (https://www.reuters.com/technology/salesforce-open-first-ai-centre-london-2024-06-06)

🛒eBay unveils new AI-tool to enhance product image backdrops

The tool, currently available for iOS users in the UK, USA, and Germany, automatically removes image backgrounds and allows users to replace them with AI-generated backdrops. The AI feature powered by Stable Diffusion will enable sellers to upload high-quality photos easily. (https://techcrunch.com/2024/06/05/ebay-ai-powered-background-enhancement-tool)

👥Asana introduces AI teammates

The platform leverages historical data based on parameters like skill sets and missing information, allowing teams to define tasks and get an overview of who is assigned which responsibility. The tool allows users to interact with it using an AI chat interface. (https://www.theverge.com/2024/6/5/24170480/asana-ai-teammate-workflow-assistant-chatbot)

🚀Writer launches a suite of No-Code development tools 

The Gen AI platform announced major expansions that included Writer Framework – a drag-and-drop tool that enables developers to connect pre-built UI components with generative AI to create custom AI applications. The tool additionally provides full flexibility to the underlying code. (https://venturebeat.com/ai/writer-launches-no-code-platform-and-framework-for-custom-enterprise-ai-applications)

A  Daily chronicle of AI Innovations June 05th 2024:

👑 Apple could crown OpenAI as King of the Valley

🚨 AI researchers demand ‘Right to Warn’

🤖 New ‘AI teammate’ can tell people what to do at work

🚗 Elon Musk explains reasoning for Nvidia chip re-route from Tesla to X

⚛️ OpenAI seeks ‘vast quantities’ of nuclear fusion energy to power superhuman AI

🚀 Cisco has unveiled HyperFabric AI Clusters in collaboration with NVIDIA
💻 Tesla’s AI ambitions on hold? Musk diverts chips to X & xAI
🤖OpenAI insiders raise concerns over oversight and safety

🤖 ChatGPT, Claude, and Perplexity experienced outages at the same time

🧠 Raspberry Pi 5 gets AI boost with Hailo extension module

📱 TECNO CAMON 30 series launches Ella-GPT AI assistant

❄️Snowflake empowers enterprise AI with new No-Code studio

💻 Zoom’s CEO envisions AI clones in meetings

👑 Apple could crown OpenAI as King of the Valley 

  • Apple and OpenAI are reportedly partnering to integrate ChatGPT into the iPhone’s operating system, potentially making it a core feature on every device.
  • The partnership could mark the end of Google’s competition with OpenAI for being the main AI chatbot on iPhones, putting OpenAI in front of Apple’s massive user base and enhancing its long-term influence in Silicon Valley.
  • While it’s unclear exactly how Apple will use ChatGPT, possibilities include powering Siri or offering it as a standalone app, with new AI features likely to be introduced as opt-in services during the Worldwide Developers Conference next week.
  • Source: https://gizmodo.com/apple-chatgpt-deal-crown-openai-sam-altman-king-valley-1851520594

🚨 AI researchers demand ‘Right to Warn’

Current and former employees from top AI labs, including OpenAI, Anthropic, and DeepMind just published an open letter calling for companies to expand whistleblower protections so workers can raise the alarm about potential AI dangers without fear of retaliation.

  • The ‘Right to Warn AI’ petition was crafted by current and former employees of OpenAI, Anthropic, and Google DeepMind.
  • The open letter was also endorsed by AI visionaries Yoshua Bengio, Geoffrey Hinton, and Stuart Russell.
  • The statement pushes for AI firms to agree to several principles:
    • Eliminating non-disparagement clauses concerning AI risk
    • Establishing and facilitating anonymous channels for raising concerns
    • Expanding whistleblower protections and anti-retaliation measures
  • Several researchers posted threadson their experience, with Daniel Kokotajlo revealing he quit OpenAI after ‘losing hope’ the company would act responsibly.

The AI safety discourse is reaching a boiling point, and there is clearly a major industry divide that transcends any one AI firm or researcher. The proposed principles seem reasonable and necessary — but it remains to be seen if the top AI leaders will actually listen.

Source: https://righttowarn.ai/

🤖 New ‘AI teammate’ can tell people what to do at work

  • Asana has introduced an “AI teammate” that not only assigns tasks but also fills in missing information, aiming to enhance team collaboration and overall efficiency within the platform.
  • The AI model leverages historical data and team relationships to match tasks with employees’ specific skill sets, ensuring that the right people are assigned to the right projects for optimal performance, and proactively gathering any missing information needed for tasks.
  • Asana’s AI teammate includes a chat interface that allows users to ask questions about their projects, similar to other AI tools from companies like Google, Amazon, and Microsoft, which are also focused on automating workflow tasks.
  • Source: https://www.theverge.com/2024/6/5/24170480/asana-ai-teammate-workflow-assistant-chatbot

🚗 Elon Musk explains reasoning for Nvidia chip re-route from Tesla to X

  • Elon Musk has acknowledged diverting Nvidia AI chips intended for Tesla to his other companies, xAI and social media firm X, claiming the chips would have been unused otherwise.
  • This move comes as Musk repositions Tesla from solely a car manufacturer to an AI and robotics company, with plans to significantly increase its chip procurement for AI training by the year’s end.
  • The timing of Musk’s confession could affect the upcoming vote on his substantial pay package, raising concerns among Tesla shareholders about potential resource misallocation to his other ventures.
  • Source: https://fortune.com/2024/06/05/elon-musk-tesla-xai-nvidia-ai-training-chips-cnbc/

⚛️ OpenAI seeks ‘vast quantities’ of nuclear fusion energy to power superhuman AI

  • OpenAI is looking to secure a deal with Helion Energy to obtain “vast quantities” of nuclear fusion energy to create superhuman artificial intelligence, despite the technology not yet being commercially available.
  • Helion Energy, supported by significant investments including $375 million from OpenAI’s CEO Sam Altman, aims to produce commercial-scale nuclear fusion energy by 2028.
  • Nuclear fusion is considered the “holy grail of clean energy” and is crucial for meeting the energy needs of future AI, but achieving it at scale remains highly challenging.
  • Source: https://www.independent.co.uk/tech/openai-nuclear-fusion-energy-ai-b2557064.html

Cisco has unveiled HyperFabric AI Clusters in collaboration with NVIDIA

Cisco and NVIDIA announced Cisco Nexus HyperFabric AI Clusters, an end-to-end infrastructure solution for scaling generative AI workloads in the data center. It combines Cisco’s AI-native networking with NVIDIA’s accelerated computing AI software and VAST’s data storage platform.

It is designed to simplify the deployment and management of generative AI applications for enterprise customers, providing centralized control across the entire AI infrastructure stack.

The Nexus HyperFabric AI cluster will be available for early customer trials in Q4 2024, with general availability expected shortly after.

Why does this matter?

This breakthrough solution aims to provide IT visibility and analytics across the entire AI infrastructure stack, allowing enterprises to focus on AI-driven revenue opportunities rather than spending excessive time on IT management.

Source: https://newsroom.cisco.com/c/r/newsroom/en/us/a/y2024/m06/cisco-reveals-nexus-hyperfabric-the-new-generative-ai-infrastructure-solution-with-nvidia-to-help-simplify-data-center-operations.html 

Tesla’s AI ambitions on hold? Musk diverts chips to X & xAI

Elon Musk instructed Nvidia to prioritize shipments of AI chips to X and xAI over Tesla, diverting over $500 million worth of Nvidia’s flagship H100 AI chips that were initially reserved for Tesla.

This decision could delay Tesla’s plans to significantly increase its acquisition of H100 chips from 35,000 to 85,000 by the end of 2024, a crucial part of Musk’s vision for transforming Tesla into “a leader in AI and robotics.”

Consequently, this move could frustrate Tesla investors who are counting on Musk to deliver on his promises regarding autonomous driving and Tesla’s AI capabilities.

Why does this matter?

Musk’s decision to prioritize chip shipments to xAI could give it a technological edge in the race to develop advanced generative AI models, potentially outpacing competitors like OpenAI, Google, and others.

Source: https://www.cnbc.com/2024/06/04/elon-musk-told-nvidia-to-ship-ai-chips-reserved-for-tesla-to-x-xai.html

OpenAI insiders raise concerns over oversight and safety

Open AI researchers are concerned about the lack of proper oversight, the influence of profit motives, and the suppression of whistleblowers working on advanced AI technologies. They warn of risks ranging “from the further entrenchment of existing inequalities to manipulation and misinformation, to the loss of control of autonomous AI systems potentially resulting in human extinction.”

They want AI companies to agree to four principles: refraining from enforcing non-disparagement agreements, establishing anonymous channels to raise concerns, allowing employees to share risk-related information publicly while protecting trade secrets, and not retaliating against whistleblowers.

Why does this matter?

Amid ongoing OpenAI controversies, the letter, coupled with the high-profile names endorsing it, will place even greater scrutiny on its practices and decision-making. This could pressure the company to be more transparent and accountable.

Source: https://venturebeat.com/ai/more-openai-researchers-slam-company-on-safety-call-for-right-to-warn-to-avert-human-extinction

What Else Is Happening in AI on June 05th 2024❗

🤖 ChatGPT, Claude, and Perplexity experienced outages at the same time

This unusual occurrence could indicate a systemic problem rather than individual issues, possibly signaling a broader infrastructure or internet-scale issue affecting these providers. (https://techcrunch.com/2024/06/04/ai-apocalypse-chatgpt-claude-and-perplexity-are-all-down-at-the-same-time)

🧠 Raspberry Pi 5 gets AI boost with Hailo extension module

Raspberry Pi launched a $70 AI Kit, an extension for the Raspberry Pi 5. It includes a neural network inference accelerator, the Hailo-8L, powered by Hailo’s AI chip. With it, the Raspberry Pi 5 can perform inferencing at 13 tera-ops per sec, facilitating tasks like object detection, semantic segmentation, and facial landmarking for camera applications. (https://techcrunch.com/2024/06/04/raspberry-pi-partners-with-hailo-for-its-ai-extension-kit)

📱 TECNO CAMON 30 series launches Ella-GPT AI assistant

It supports over 70 languages, helps with daily tasks and content creation, and improves user interaction with features like real-time translations, voice commands, and personalized assistance. Additional capabilities include Ask AI for text editing and grammar checks and AI Generate for turning sketches into images.  (https://www.businesstoday.in/technology/news/story/tecno-camon-30-series-introduces-new-ai-assistant-ella-gpt-432156-2024-06-05)

❄️Snowflake empowers enterprise AI with new No-Code studio

It announced several updates to its Cortex AI service and Snowflake ML. The introduction of No-Code AI & ML Studio stands out among these enhancements, enabling every enterprise user to construct AI applications tailored to their specific use cases without requiring coding expertise.  (https://venturebeat.com/ai/snowflake-doubles-down-on-enterprise-ai-with-no-code-studio-and-more)

💻 Zoom’s CEO envisions AI clones in meetings

Zoom’s CEO, Eric Yuan, envisions AI-driven digital avatars, or “digital twins,” representing humans in meetings, potentially reducing the workweek to three or four days. He argues  AI can efficiently manage tasks like Zoom calls, chats, and emails, allowing people to reclaim time spent in meetings. This initiative forms part of Zoom’s 2.0 journey, aiming to evolve beyond a mere videoconferencing tool.  (https://www.theverge.com/2024/6/3/24168733/zoom-ceo-ai-clones-digital-twins-videoconferencing-decoder-interview)

Google’s AI Overviews appear on 70% fewer Search results pages now: https://www.zdnet.com/article/googles-ai-overviews-appear-on-70-fewer-search-results-pages-now

OpenAI CEO Sam Altman’s chip plans were ‘too aggressive’ for TSMC: https://the-decoder.com/openai-ceo-sam-altmans-chip-plans-were-too-aggressive-for-tsmc/

AI startup Cohere gets $450 million from Nvidia, Salesforce, and others: https://the-decoder.com/ai-startup-cohere-gets-450-million-from-nvidia-salesforce-and-others/

Cisco announces $1bn AI investment fund: https://techmonitor.ai/technology/ai-and-automation/cisco-ai-invest-fund

MIT’s new AI chatbot can simulate users’ future selves: https://www.newsbytesapp.com/news/science/mit-ai-researchers-create-chatbot-to-simulate-future/story

📱 AI Training: Explain images in real-time with GPT-4o

This hack lets you access GPT-4o directly from an iPhone via a custom shortcut, allowing you to analyze and explain images in real time.

  1. Head over to OpenAI’s platform to obtain an API key. Make sure you have some credits in your account.
  2. Go to your iPhone Shortcuts and click the “+” button.
  3. Add different actions before your OpenAI API call, e.g., your desired prompt as “Text”, your API key as “Text”, and “Take a screenshot”.
  4. Call GPT-4o by using “*https://api.openai.com/v1/chat/completions”* and obtain a description of what you are seeing!

📶 AI Research: OpenAI researcher: AGI coming by 2027

Former OpenAI researcher Leopold Aschenbrenner just released a new essay series detailing his view on AGI, saying the rate of AI progress will be the most intense and volatile events in human history.

  • Aschenbrenner says that ‘nobody is pricing in’ what is coming in AI, and to expect another GPT-2 to GPT-4 level jump by 2027 (that would take us to AGI).
  • The researcher predicts that hundreds of millions of AGI would then rapidly accelerate progress, compressing decades of progress into a year.
  • He also discussed the economic and military advantages that will come with AGI, calling it a national security issue that needs its own ‘Manhattan Project’.
  • Aschenbrenner reiterated these views on the Dwarkesh Podcast, also revealing he was fired from OpenAI after raising AI security concerns.

As an insider at OpenAI, Aschenbrenner’s analysis carries weight — and his projections paint a striking picture of how radically AGI will reshape the world. His commentary on the firing also raises even more drama, given OpenAI’s current media battle with former researchers and the board over safety concerns.

Source: https://situational-awareness.ai/wp-content/uploads/2024/06/situationalawareness.pdf

New AI Jobs Opportunities on June 05th 2024:

  • 🎨 Siena – Head of Design: https://jobs.therundown.ai/jobs/61293280-head-of-design
  • ⚙️ Figure AI – Software Engineer, Developer Tools and Productivity: https://jobs.therundown.ai/jobs/59968286-software-engineer-developer-tools-and-productivity
  • 🧠 Snorkel – ML Enablement Lead: https://jobs.therundown.ai/jobs/61783157-ml-enablement-lead
  • 📐 Cohere – Senior Manager, Solutions Architect: https://jobs.therundown.ai/jobs/61771861-senior-manager-solutions-architect

A  Daily chronicle of AI Innovations June 04th 2024:

💻 Intel’s new data center chips handle demanding AI workloads
📦 Amazon’s Project PI detects defective products before shipping
☁️ Microsoft’s Aurora AI could transform weather forecasting

🎓 High-quality education data key to AI performance: Research

👎 LeCun criticizes Musk for mistreating scientists and spreading misinformation 

💰 Microsoft to invest $3.2 billion in Sweden to expand AI and cloud infrastructure

🤖 Microsoft identifies few AI deep fakes in the EU election

☕️ Robots serve up coffee at Starbucks

📦 Amazon’s AI ‘private investigator’

💻 Intel’s new data center chips handle demanding AI workloads

Intel has announced next-generation Xeon 6 server processors to regain the data center market share it had been losing to AMD. They come in two varieties. The larger, more powerful version is designed to run the computations necessary to generate responses from complex AI models and other tasks requiring increased horsepower. Intel plans to help companies modernize their aging data center systems with Xeon 6 chips so they can generate new digital capabilities.

Intel also revealed that its Gaudi 3 AI accelerator chips would be priced much lower than its rivals’ products.

Why does it matter?

As more companies have started to deploy AI apps and models, the AI hardware space is getting heated with competition. Intel seems to be one of the only companies innovating across the full spectrum of the AI market opportunity– from semiconductor manufacturing to PCs and data center systems.

Source: https://venturebeat.com/ai/intel-reveals-xeon-6-processor-enterprise-ai-gaudi-3-accelerator-price 

📦 Amazon’s Project PI detects defective products before shipping

Amazon has launched Project PI, which uses AI to scan products for defects before shipping them to customers. This AI system combines computer vision to visually inspect items with generative AI models that can understand things like text on packages.

As products go through a scanning tunnel, the AI checks for damage, incorrect colors/sizes, or expired dates. If it finds a problem, that item is isolated to evaluate the defect.  Project PI already operates in several of Amazon’s warehouses across North America. The system catches millions of defective products daily before they reach customers.

Why does it matter?

Using innovative AI systems, retailers can avoid dealing with returns and reshipments, reducing costs and inefficiencies. By cutting down unnecessary shipping, retailers minimize environmental impact and carbon emissions, contributing to sustainability goals.

Source: https://www.theverge.com/2024/6/3/24170567/amazons-project-pi-product-defect-return-ai-computer-vision

☁️ Microsoft’s Aurora AI could transform weather forecasting

Microsoft has developed a powerful new AI foundation model called Aurora that can make highly accurate weather predictions. It is trained on over a million diverse weather and climate data hours. This allows it to develop a comprehensive understanding of atmospheric dynamics and excel at forecasting various weather variables like temperature, wind speed, air pollution levels, and greenhouse gas concentrations.

What sets Aurora apart is its ability to capture intricate details at high spatial resolution (around 11km) while being much faster and more computationally efficient than traditional numerical weather prediction systems. Aurora’s flexible architecture and training on heterogeneous datasets enable it to adapt to different forecasting tasks and resolutions.

Why does it matter?

This major advancement in AI-based weather forecasting could help communities prepare for extreme weather events like storms. AI will also play a bigger role in predicting the impacts of climate change. We may be nearing days when weatherman’s predictions will be 100% accurate.

Source: https://www.microsoft.com/en-us/research/blog/introducing-aurora-the-first-large-scale-foundation-model-of-the-atmosphere

☕️ Robots serve up coffee at Starbucks

South Korean search giant Naver shared a video of the company’s in-office autonomous Starbucks location — with 100 robots delivering coffee and other items throughout the building.

  • Naver’s autonomous wheeled “Rookie” robots navigate the building’s 36 floors to bring packages, coffee, and lunch to employees.
  • Rookie bots are assisted by dual-armed “Ambidex” robots, which are lightweight and dexterous for safer human interactions.
  • Both robots connect to the company’s ARC system, which enables control of navigation, planning, and processing for the entire fleet through cloud computing.
  • Naver also developed RoboPort, a dedicated elevator system that allows robots to move quickly and efficiently between floors.

Naver’s putting its robots to work in the ‘real world’ — and while the location is just a testing ground for now, it’s also a glimpse into how service robotics will be integrated into the mainstream sooner than many expect.

📦 Amazon’s AI ‘private investigator’

Amazon just unveiled Project P.I., an AI system that scans products in the company’s fulfillment centers to detect damaged or incorrect items before they ship to reduce returns and waste.

  • Project P.I. uses AI and computer vision to spot defects like damaged products as well as wrong colors/sizes as items pass through imaging tunnels.
  • The system is already in place across the company’s North American fulfillment centers, with plans to expand globally throughout the year.
  • Amazon also utilizes a multimodal LLM to investigate issues further, combining customer feedback with Project P.I. images to identify the source problem.

Amazon’s no stranger to adopting AI across its operations, and Project P.I. is yet another innovative approach to supercharge efficiency. Soon, the company will likely combine these talents with advanced warehouse robots — taking humans out of the fulfillment process altogether.

Source: https://www.aboutamazon.com/news/innovation-at-amazon/amazon-ai-sustainability-carbon-footprint-product-defects

⛈️ AI RESEARCH: Microsoft’s AI weather forecasting model

Microsoft researchers just introduced Aurora, a new AI foundation model trained on over 1M hours of weather and climate data that can generate accurate weather forecasting insights.

  • Aurora produces accurate forecasts across a variety of weather scenarios, including extreme events or areas with limited data.
  • The 1.3B parameter model can generate a 5-day global air pollution prediction in under 60 seconds.
  • Aurora also produces 10-day global forecasts at high resolutions, beating both the top models and specialized AI weather systems.

 Aurora’s success could usher in a huge shift in the way we do weather forecasting. With the recent progress of AI models, combined with massive amounts of atmospheric data — the day that the weatherman is truly right 100% of the time might be coming sooner than most people probably think.

Source: https://www.microsoft.com/en-us/research/blog/introducing-aurora-the-first-large-scale-foundation-model-of-the-atmosphere

New AI Job Opportunities on June 04th 2024

🎨 Adept AI – Product Designer: https://jobs.therundown.ai/jobs/61589310-product-designer
⚙️ Luma AI – Senior Distributed Systems Engineer: https://jobs.therundown.ai/jobs/61614624-senior-distributed-systems-engineer
🚗 Scale AI – Account Executive, Automotive: https://jobs.therundown.ai/jobs/61614640-account-executive-automotive
🧑‍💻 Lambda – Technical Program Manager: https://jobs.therundown.ai/jobs/61589251-technical-program-manager

What Else Is Happening in AI on June 04th 2024❗

🚫 Hugging Face detects ‘unauthorized access’ to its AI model hosting platform

Last week, Hugging Face detected unauthorized access to Spaces, its platform for creating, sharing, and hosting AI models. It suspects some Spaces secrets (private pieces of info that act as keys to unlock protected resources like accounts, tools, and dev environments) have leaked. It has taken steps to remediate this. (https://huggingface.co/blog/space-secrets-disclosure)

🎓 High-quality education data key to AI performance: Research

Researchers created a high-quality dataset called FineWeb-Edu by filtering an existing web dataset for educational content. Language models trained on FineWeb-Edu significantly outperformed models trained on unfiltered datasets. The research shows that data quality and diversity are more important than dataset size for training effective AI models. (https://x.com/gui_penedo/status/1797173053123916036)

👎 LeCun criticizes Musk for mistreating scientists and spreading misinformation 

LeCun has again rebuked Musk on X after they had a heated feud on X last week. This time, LeCun accused Musk of forcing researchers to work in secrecy instead of allowing them to publish their work, which slows scientific progress. He also accused Musk of falsely predicting AI and autonomous vehicles. (https://venturebeat.com/ai/yann-lecun-ai-pioneer-sharply-criticizes-elon-musk-over-treatment-of-scientists-and-spreading-of-misinformation)

💰 Microsoft to invest $3.2 billion in Sweden to expand AI and cloud infrastructure

Microsoft will invest $3.2 billion over two years to expand its cloud and AI infrastructure in Sweden. Microsoft’s biggest investment to date in Sweden includes a pledge to help train some 250,000 people with AI skills, corresponding to 2.4% of the population, which will help boost the Nordic country’s competitiveness in generative AI. (https://www.reuters.com/technology/microsoft-invest-32-bln-swedish-cloud-ai-2024-06-03)

🤖 Microsoft identifies few AI deep fakes in the EU election

As the European Union prepares for its elections in June 2024, the threat of AI-generated deepfakes has become a significant concern. Microsoft President Brad Smith highlighted this burning issue, emphasizing its potential impact on the democratic process and the steps Microsoft is taking to mitigate these risks. (https://www.reuters.com/technology/few-ai-deepfakes-identified-eu-elections-microsoft-president-says-2024-06-03 )

A  Daily chronicle of AI Innovations June 01-03 2024:

🤯 Windows AI feature is a security ‘disaster’

👥 Zoom CEO wants AI clones in meetings

🔧 Nvidia and AMD unveil next generation AI chips as competition heats up

🎮 Nvidia announces an AI gaming assistant

📢 Nvidia CEO drops a series of AI announcements
🚀 AMD outlined new chip architecture strategy for AI data centers
🔊 ElevenLabs’ Text to Sound AI wows creators

🤯 Windows AI feature is a security ‘disaster’

  • Microsoft’s new AI-powered Recall feature, which captures and stores screenshots of all user activity on a PC, has been criticized as a cybersecurity “disaster” due to significant security vulnerabilities discovered during testing.
  • Cybersecurity expert Kevin Beaumont found that Recall stores screenshots in an unencrypted plain text database, making it easy for malware and attackers to access sensitive data, despite Microsoft’s claims of a secure, encrypted experience.
  • Privacy advocates and the UK’s ICC have raised concerns about Recall’s potential to expose personal information, with criticisms focusing on its default enabled status and the lack of comprehensive content moderation to protect sensitive data like passwords and financial information.
  • Source: https://www.theverge.com/2024/6/3/24170305/microsoft-windows-recall-ai-screenshots-security-privacy-issues

👥 Zoom CEO wants AI clones in meetings

  • Zoom CEO Eric Yuan envisions using AI-powered digital twins to attend meetings on behalf of users, allowing people to manage their time more effectively.
  • Yuan aims to expand Zoom beyond videoconferencing into the broader enterprise software market, competing with giants like Microsoft and Google by integrating productivity tools such as email and chat.
  • Zoom is heavily investing in AI technology to facilitate this transition, although Yuan acknowledges challenges like AI hallucination that need to be addressed before fully realizing his vision.
  • Source: https://www.theverge.com/2024/6/3/24168733/zoom-ceo-ai-clones-digital-twins-videoconferencing-decoder-interview

🔧 Nvidia and AMD unveil next generation AI chips as competition heats up

  • Nvidia and AMD showcased their latest AI chips at a major computing conference, heightening their rivalry in the AI chip market.
  • Nvidia CEO Jensen Huang emphasized his company’s dominance and highlighted a new chip, Rubin, set for release in 2026, while AMD CEO Lisa Su focused on teamwork and collaborative presentations with industry partners.
  • The competition between the two companies is significant for AI development, with Nvidia promoting a proprietary ecosystem and AMD advocating for open standards and interoperability.
  • Source: https://finance.yahoo.com/news/nvidia-amd-chiefs-square-off-101007730.html

🎮 Nvidia announces an AI gaming assistant

  • Nvidia recently demonstrated its AI assistant, G-Assist, showcasing its evolution from an April Fools’ prank to a powerful tool for game developers and RTX GPU owners.
  • G-Assist can respond to voice queries, understand in-game activities, provide guidance based on player’s skill points, and offer performance-enhancing recommendations for a smooth gaming experience.
  • The AI assistant optimizes PC settings, monitors latency and frame rates, detects refresh rate discrepancies, and suggests ways to boost performance, including overclocking the GPU if necessary.
  • Source: https://www.newsbytesapp.com/news/science/nvidia-g-assist-ai-assistant-breaks-cover-features-specifications/story

Nvidia CEO drops a series of AI announcements

Nvidia CEO Jensen Huang revealed the company’s ambitious plans for annual AI accelerator upgrades, targeting a broader range of industries to expand its customer base.

  • It will release the Blackwell Ultra chip in 2025 and the next-generation Rubin platform in 2026.
  • It is also releasing a new server design, MGX, to help companies like HPE and Dell bring products to market faster.
  • They are promoting the use of digital twins in its Omniverse virtual world, showcasing a digital twin of Earth for sophisticated modeling tasks.
  • Introduces Project G-Assist, an RTX-powered AI assistant technology that provides context-aware help for PC games and apps.
  • G-Assist uses voice or text inputs and game window snapshots to provide personalized responses based on in-game context.
  • Developers can customize the AI models for specific games or apps, and they can run on the cloud or locally on GeForce RTX AI PCs and laptops.
  • Nvidia partnered with Studio Wildcard for a tech demo using ARK: Survival Ascended, showcasing how G-Assist can help with quests, items, lore, and challenging bosses. Check out full keynote speech:
  • https://www.youtube.com/live/pKXDVsWZmUU?si=mOPV2g0EP_gcUfMa

Why does it matter?

These announcements show how eager Nvidia is to retain its position as a leader in the AI hardware market. In addition to pushing the acceleration of AI chips, Nvidia is developing new tools to shape AI’s implementation in multiple sectors.

Source: https://www.bloomberg.com/news/articles/2024-06-02/jensen-huang-computex-keynote-nvidia-reveals-new-ai-software-and-services

AMD outlined new chip architecture strategy for AI data centers 

AMD CEO Lisa Su introduced new AI processors at Computex, including the MI325X accelerator, set to be available in Q4 2024.

  • The CEO announced the MI325X accelerator, which will be released in Q4 2024, and outlined the company’s plan to develop AI chips over the next two years.
  • Introduced the MI350 series, expected in 2025, which promises a 35x improvement in inference performance compared to the current MI300 series.
  • The company also teased the MI400 series, slated for 2026, based on the mysterious “Next” architecture.

With AMD and Nvidia moving to annual release cycles, the competition is heating up to meet the soaring demand for AI semiconductors.

Why does it matter?

AMD’s aggressive push to challenge Nvidia’s market leadership could lead to increased innovation, lower prices, and more widespread adoption of AI across various sectors.

Source: https://www.reuters.com/technology/amd-launches-new-ai-chips-take-leader-nvidia-2024-06-03

ElevenLabs’ Text to Sound AI wows creators

ElevenLabs introduces Text to Sound, an AI model that generates sound effects, instrumental tracks, soundscapes, and character voices from text prompts. The tool aims to help film, TV, video games, and social media creators produce high-quality audio content quickly and affordably.

They have partnered with Shutterstock to fine-tune the model using their diverse audio library of licensed tracks. Users can generate sound effects by logging in, describing the desired sound, and downloading the best results.

Note: This tool doesn’t have a content filter and can generate any raw content through conditional prompting.

Why does it matter?

It could significantly reduce production costs and timelines by simplifying the development of high-quality sound effects, music, and voices, encouraging smaller studios and individual creators to compete with larger players.

Source: https://elevenlabs.io/blog/sound-effects-are-here

What Else Is Happening in AI on June 03rd 2024❗

🤖 xAI is developing two new modes for the Grok AI chatbot

These two new modes are ‘Socrates’ and ‘DEI’ (Diversity, Equity and Inclusion). Grok currently offers normal mode, fun mode, and the recently announced ‘Unhinged’ mode. In DEI mode, Grok will act as a ‘Senior VP of Diversity and Inclusion’ and provide correct answers even to incorrect questions. The function of the ‘Socrates’ mode is not yet clear. The release date for these new modes has not been confirmed. (https://www.latestly.com/socially/technology/grok-new-modes-elon-musks-xai-working-on-two-new-modes-called-socrates-and-dei-likely-to-be-announced-soon-6006912.html)

🎬 Sony Pictures CEOannounces AI plans for better movie and TV production

The move comes amid negotiations between Hollywood’s major crew union and top studios. While union agreements will define the extent of AI use in the industry, the tech is already eliminating jobs in Hollywood, particularly in voice acting, concept art, VFX, and postproduction. (https://www.hollywoodreporter.com/business/business-news/sony-pictures-adopt-ai-streamline-production-says-ceo-tony-vinciquerra-1235912109)

🚀 Arm predicts 100 Billion devices AI-ready by 2025

According to CEO Rene Haas at the Computex forum in Taipei, Arm Holdings expects a staggering 100 billion Arm devices worldwide to be primed for artificial intelligence by the end of 2025. (https://www.reuters.com/technology/arm-expects-100-billion-arm-devices-will-be-ready-ai-by-end-2025-2024-06-03)

💬 Meta’s AI generates bizarre summaries of Facebook comments on various posts

The AI picks up on both serious and lighthearted comments, often highlighting the more outlandish ones. It’s unclear how Meta chooses which posts to display these summaries on. However, the summaries have raised privacy concerns, as Meta feeds user comments into its AI system. (https://www.theverge.com/2024/5/31/24168802/meta-ai-facebook-comments-summaries)

🗑️ Finnish startup Binit is developing an AI household waste tracker

The gadget, designed to be mounted in the kitchen, has cameras and sensors to scan items before throwing them away. Binit uses OpenAI’s GPT for image recognition, achieving nearly 98% accuracy in trash recognition. The app provides analytics, feedback, and gamification to encourage users to reduce waste, with tests showing a 40% reduction in mixed bin waste. (https://techcrunch.com/2024/06/03/binit-is-bringing-ai-to-trash)

Latest AI Tools on June 03rd  2024:

  • 👩‍🎤 Consistent Character Model – Create consistent character images in different poses: https://supertools.therundown.ai/content/consistent-character-mode
  • ✍️ Perplexity Pages – Turn your research into shareable articles with AI: https://supertools.therundown.ai/content/paper-clipserplexity-ai
  • 🎥 PixVerse Magic Brush – Brush areas, mark directions, and watch images come to life: https://supertools.therundown.ai/content/pixverse
  • 🔰 Glaze – Protect artists from generative AI: https://supertools.therundown.ai/content/glaze
  • 🎨 Fontjoy – Helps designers find perfect font pairings in one click: https://supertools.therundown.ai/content/fontjoy
  • 🤗 Eva Coach – An AI assistant to help manage work-related stress: https://supertools.therundown.ai/content/eve-coach

Latest AI Jobs on June 03rd 2024:

  • 🔧 Figure AI – Solutions Lead: https://jobs.therundown.ai/jobs/59875235-solutions-lead
  • 💻 Kumo – Software Engineer Internship: https://jobs.therundown.ai/jobs/61517389-software-engineer-internship
  • 🛠️ Palantir Technologies – Backend Software Engineer, Application Development: https://jobs.therundown.ai/jobs/61517369-backend-software-engineer-application-development
  • 📖 Meta – Technical Program Manager, AI Research: https://jobs.therundown.ai/jobs/60072576-technical-program-manager-ai-research

A Daily Chronicle of AI Innovations in May 2024

  • AI: The Ultimate Sherlocking?
    by /u/mintone (Artificial Intelligence) on July 26, 2024 at 12:16 pm

    submitted by /u/mintone [link] [comments]

  • Speech-to-Text Solution for Multilingual Sentences / Mixed-language speech
    by /u/simbaninja33 (Artificial Intelligence Gateway) on July 26, 2024 at 11:54 am

    I am looking for a speech-to-text solution, either paid or open-source, that can accurately transcribe speech containing a mix of two languages within the same sentence. I have explored options like Microsoft Azure, Google Cloud, and OpenAI, but haven't found a satisfactory solution yet. For example, I need the solution to handle sentences like: "I have tried the restaurant yesterday, it is muy muy bueno, they serve some of the pizza, que haria mi abuela super celoza de la receta." "I went to the store y compré un poco de pan because we were running low." I have already tried Microsoft Azure, which can handle multiple languages, but only when they are not mixed within the same sentence (as mentioned in their documentation). Google Cloud's speech-to-text fails to accurately transcribe mixed-language speech, and OpenAI doesn't seem to offer this functionality. I am open to both continuous real-time speech recognition and file-based recognition. For real-time applications, I am also willing to consider workarounds, such as implementing a "button" that can be clicked to quickly switch between the main language and the second language. If anyone has experience with a solution that can handle this type of mixed-language speech recognition, I would greatly appreciate any suggestions or recommendations. Thank you in advance for your help! submitted by /u/simbaninja33 [link] [comments]

  • Any open source AI model with web search abilities?
    by /u/david8840 (Artificial Intelligence Gateway) on July 26, 2024 at 11:45 am

    Is there any open source AI model with web search abilities? I want to be able to ask it questions which require real time internet searching, for example "What is the weather like now in NY?" submitted by /u/david8840 [link] [comments]

  • Which companies are leading the way in AI detection? (for audio/video deepfakes, etc.?)
    by /u/ProfessionalHat3555 (Artificial Intelligence Gateway) on July 26, 2024 at 11:21 am

    So I was listening to the most recent Bill Simmons pod w/ Derek Thompson where they discuss conspiracy theories and AI shit-detection (40:00-48:00 if you're curious)... 1ST Q: what companies are you aware of that are already working on AI detection? 2ND Q: where do you think the AI detection slice of the market is going? Will there be consumer-grade products that we can use to run, say, a political video through a detection software & get a % of realness rating on it? Will these tools ONLY be available to big conglomerates who become the purveyors of truth? 3RD Q: If we're UNABLE to do this at-scale yet, what would need to happen tech-wise for AI detection to become more accessible to more people? (disclaimer: I'm not a dev) submitted by /u/ProfessionalHat3555 [link] [comments]

  • AI can't take people's jobs if there's no people.
    by /u/baalzimon (Artificial Intelligence Gateway) on July 26, 2024 at 10:53 am

    Looks more and more likely that human populations will decline in the future. Maybe the workforce will just be AI robots rather than young people. PEW: The Experiences of U.S. Adults Who Don’t Have Children 57% of adults under 50 who say they’re unlikely to ever have kids say a major reason is they just don’t want to; 31% of those ages 50 and older without kids cite this as a reason they never had them https://www.pewresearch.org/social-trends/2024/07/25/the-experiences-of-u-s-adults-who-dont-have-children/ submitted by /u/baalzimon [link] [comments]

  • UK School Under Fire for Unlawful Facial-Recognition Use
    by /u/Think_Cat1101 (Artificial Intelligence Gateway) on July 26, 2024 at 10:43 am

    https://www.msn.com/en-us/news/technology/uk-school-under-fire-for-unlawful-facial-recognition-use/ar-BB1qEmeX?cvid=6dfe65854c6e4c2ad473b0e649e795b2&ei=10 submitted by /u/Think_Cat1101 [link] [comments]

  • OpenAI reveals 'SearchGPT'
    by /u/Mindful-AI (Artificial Intelligence Gateway) on July 26, 2024 at 10:41 am

    submitted by /u/Mindful-AI [link] [comments]

  • Amazon’s AI Chip Revolution: How They’re Ditching Nvidia’s High Prices and Speeding Ahead
    by /u/alyis4u (Artificial Intelligence Gateway) on July 26, 2024 at 9:23 am

    Six engineers tested a brand-new, secret server design on a Friday afternoon in Amazon.com’s chip lab in Austin, Texas. Amazon executive Rami Sinno said on Friday during a visit to the lab that the server was full of Amazon’s AI chips, which compete with Nvidia’s chips and are the market leader.https://theaiwired.com/amazons-ai-chip-revolution-how-theyre-ditching-nvidias-high-prices-and-speeding-ahead/ submitted by /u/alyis4u [link] [comments]

  • OpenAI's SearchGPT Is Coming For Google Search; Here Are The Features That Will Reportedly Make It Better
    by /u/vinaylovestotravel (Artificial Intelligence Gateway) on July 26, 2024 at 9:00 am

    Dubbed "SearchGPT," the tool will offer "fast and timely answers with clear and relevant sources" by referencing content from websites and news publishers, including OpenAI content partners such as News Corp (The Post's parent company) and The Atlantic. Read more: https://www.ibtimes.co.uk/openais-searchgpt-coming-google-search-here-are-features-that-will-reportedly-make-it-better-1725770 submitted by /u/vinaylovestotravel [link] [comments]

  • Deleting chats from Blackbox AI?
    by /u/Intelligent-Fig-7791 (Artificial Intelligence Gateway) on July 26, 2024 at 7:40 am

    How on earth do you delete chats from blackbox.ai ? it seems like all chats are public by default submitted by /u/Intelligent-Fig-7791 [link] [comments]

A Daily Chronicle of AI Innovations in May 2024

AI Innovations in May 2024

AI Innovations in May 2024.

Welcome to our blog series “AI Innovations in May 2024”! This is an evolving article that will be updated daily throughout the month of May 2024 to bring you the latest news and developments in the world of artificial intelligence. As we move further into the 21st century, the pace of AI innovation continues to accelerate at a breathtaking rate. Each day, researchers, engineers, and visionaries are pushing the boundaries of what’s possible, unlocking new capabilities and applications that are transforming industries, enhancing our lives, and shaping the future. In this blog, we’ll dive deep into the most exciting AI breakthroughs, advancements, and milestones happening in May 2024. From groundbreaking AI-powered technologies and cutting-edge research to the societal and ethical implications of these innovations, we’ll provide you with a comprehensive and insightful look at the rapidly evolving world of artificial intelligence. Whether you’re an AI enthusiast, a tech-savvy professional, or simply someone curious about the future, this blog will keep you informed, inspired, and engaged. So, join us on this journey of discovery as we explore the frontiers of AI and uncover the innovations that are shaping our world. Stay tuned for daily updates, and get ready to be amazed by the incredible advancements that are happening in the world of AI!

LISTEN DAILY AT OUR PODCAST HERE

Experience the transformative capabilities of AI with “Read Aloud For Me – AI Dashboard – AI Tools Catalog – AI Tools Recommender” – your ultimate AI Dashboard and Hub. Seamlessly access a comprehensive suite of top-tier AI tools within a single app, meticulously crafted to enhance your efficiency and streamline your digital interactions. Now available on the web at readaloudforme.com and across popular app platforms including Apple, Google, and Microsoft, “Read Aloud For Me – AI Dashboard” places the future of AI at your fingertips, blending convenience with cutting-edge innovation. Whether for professional endeavors, educational pursuits, or personal enrichment, our app serves as your portal to the forefront of AI technologies. Embrace the future today by downloading our app and revolutionize your engagement with AI tools.

AI Dashboard - Wonderland Bedtime Stories - Safe Ai for All
AI Dashboard – Wonderland Bedtime Stories – Safe Ai for All

A  Daily chronicle of AI Innovations May 31st 2024:

⚠️ OpenAI reports misuse of its AI for ‘deceptive activity’
🏫 OpenAI introduces ChatGPT Edu for universities
💼 Tech giants form industry group to develop next-gen AI chip

💰 Google mistakenly erased $135 billion from customer account

📱 Apple planning AI-driven overhaul for Siri to manage individual apps

🇺🇸 TikTok to create a US-only algorithm

🦷 World’s first tooth-regrowing drug approved for human trials

📚 Perplexity launches AI publishing platform

🎓 OpenAI launches ChatGPT Edu for universities

⚠️ OpenAI reports misuse of its AI for ‘deceptive activity’

OpenAI identified and disrupted five covert influence operations that used its generative AI models for “deceptive activity” across the internet.

OpenAI said the threat actors used its AI models to generate short comments, longer articles in various languages, made-up names, and bios for social media accounts over the last three months. These campaigns included threat actors from Russia, China, Iran, and Israel. It focused on issues including Russia’s invasion of Ukraine, the conflict in Gaza, the Indian elections, and politics in Europe and the United States, among others.

OpenAI has also published a trend analysis report that describes the behavior of these malicious actors in detail.

Why does it matter?

This provides concrete evidence of generative AI being used to spread misinformation and manipulate global affairs. While there were no novel attacks this time, defending against them in the future will require continued investment in monitoring, safeguards, and public awareness from AI developers and platforms.

Source: https://openai.com/index/openai-and-reddit-partnership

💼 Tech giants form industry group to develop next-gen AI chip

Intel, Google, Microsoft, Meta, and other tech heavyweights are establishing a new industry group called the Ultra Accelerator Link (UALink) Promoter Group to guide the development of the components that link together AI accelerator chips in data centers. The group also counts AMD, Hewlett Packard Enterprise, Broadcom, and Cisco among its members.

It is proposing a new industry standard to connect the AI accelerator chips found within a growing number of servers. It’s version one, UALink 1.0, will connect up to 1,024 AI accelerators — GPUs only — across a single computing “pod.” It is set to arrive in Q4 2024.

Why does it matter?

Nvidia, currently the biggest player in the AI chip market with an 80% share, is not in the group. Tech giants have been keen to reduce their dependence on Nvidia by working on their own custom chips. This seems to be the latest effort to break Nvidia’s dominance.


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

Source: https://techcrunch.com/2024/05/30/tech-giants-form-new-group-in-effort-to-wean-off-of-nvidia-hardware

📱 Apple planning AI-driven overhaul for Siri to manage individual apps

  • Apple is planning a major AI upgrade for Siri in iOS 18, enabling users to control specific features within iPhone apps using voice commands, initially limited to Apple-developed apps.
  • The update will eventually allow Siri to handle multiple tasks in a single request, such as summarizing a recorded meeting and sending it to a friend, as part of Apple’s larger AI strategy to be unveiled at WWDC on June 10.
  • New AI-driven capabilities will enable Siri to control iPhones more precisely, including opening files, moving notes, managing emails, and summarizing articles, enhancing device navigation and functionality.
  • Source: https://www.newsbytesapp.com/news/science/siri-to-gain-enhanced-app-control-in-ios-18/story

📚 Perplexity launches AI publishing platform

  • Perplexity introduced ‘Pages’, an AI-driven feature that creates customizable webpages for research and writing based on user prompts.
  • The ‘Pages’ tool gathers information using Perplexity’s AI models, organizes it into sections with citations, and tailors content for different audience levels but requires new prompts to correct errors.
  • The Verge tested ‘Pages’ on a complex topic and noted its difficulty with intricate subjects, showing it might struggle with in-depth research despite effectively explaining fundamental concepts.
  • Source: https://www.newsbytesapp.com/news/science/perplexity-launches-new-ai-feature-called-pages/story

🎓 OpenAI launches ChatGPT Edu for universities

  • OpenAI has introduced ChatGPT Edu, a new version of ChatGPT, aimed at improving the academic and operational efficiency of universities with advanced text interpretation and data analysis capabilities, and support for over 50 languages.
  • This version was created following the success of ChatGPT Enterprise in academic institutions such as Oxford, Wharton, and ASU, where it has been used for tasks ranging from large dataset analysis to personalized language practice.
  • ChatGPT Edu provides affordable access to the GPT-4o model, enhanced message limits, robust security measures, and assures that data and conversations will not be used for training OpenAI models, ensuring a safe environment for educational use.
  • Source: https://www.maginative.com/article/openai-launches-chatgpt-edu-for-universities/

OpenAI says it stopped multiple covert influence operations that abused its AI models. https://www.engadget.com/openai-says-it-stopped-multiple-covert-influence-operations-that-abused-its-ai-models-225115466.html

Google says it fixed the AI Overviews everyone has been roasting for weeks. https://bgr.com/tech/google-says-it-fixed-ai-overviews-blames-some-of-the-problems-on-you/

Anthropic’s Claude AI now autonomously interacts with external data and tools. https://venturebeat.com/ai/anthropic-claude-ai-now-autonomously-interacts-with-external-data-and-tools/

🎬 The Simulation unveils ‘Netflix of AI’

AI entertainment startup The Simulation (formerly Fable Studio) just launched Showrunner, a platform that allows users to generate and watch AI-powered TV shows set in virtual simulated worlds.

  • Showrunner combines multi-agent simulations with LLMs to create interactive content aligned with specific virtual worlds.
  • Users can watch, direct, and star in shows set in virtual environments populated by AI characters, with the ability to craft episodes from prompts.
  • The platform is launching with 10 original shows, with users able to generate new episodes and edit deeper into scripts, shots, and voices.
  • Showrunner is opening in alpha to a limited number of users, with select user-created episodes receiving payment, rev share, and IMDB credits.
  • The studio went viral last year after the release of its South Park episode demos and research paper on its SHOW-1 model.

Showrunner is a wild step towards the merging of AI, gaming, and traditional entertainment, enabling users to not only consume but create stories within simulated worlds. The lines are blurring fast between creators and audiences — and the traditional Hollywood media model may never be the same.

Source: https://x.com/fablesimulation/status/1796245808066740274

New AI Job Opportunities on May 31st 2024

  • 👁️ Waymo – Software Engineer, Computer Vision/Deep Learning: https://jobs.therundown.ai/jobs/60918768-software-engineer-computer-vision-deep-learning
  • 📈 Notable – Strategic Account Executive
  • 💼 Findem – Director of Sales, Enterprise
  • 💻 Twelve Labs – Lead Software Engineer, Frontend

What Else Is Happening in AI on May 31st 2024❗

🔗Anthropic’s Claude can now autonomously interact with external data and tools

Anthropic has announced the general availability of Tool Use for its AI, Claude. It allows Claude to autonomously interact with external data sources, APIs, and tools, making it a business-ready trusted AI solution that can be confidently adopted by enterprises across various industries. (https://venturebeat.com/ai/anthropic-claude-ai-now-autonomously-interacts-with-external-data-and-tools)

✨Perplexity goes beyond AI search, launches publishing platform ‘Pages’

It is a tool to transform disorganized AI knowledge and research into easily digestible articles and reports for sharing. Users just have to describe the topic, select the target audience, and it will produce an in-depth article with a clear title, subheadings, associated media, and relevant citations. (https://x.com/perplexity_ai/status/1796203494401040846)

🔍Google poaches AWS and Microsoft executives to strengthen AI offerings for cloud customers

Google hired two executives from these rivals as it builds out its AI ranks with the goal of boosting its cloud business. Saurabh Tiwary, a former corporate VP at Mircosoft, will join a newly created role as general manager of cloud AI. Raj Pai, a former VP at AWS, will oversee product management of the cloud AI team. (https://www.theinformation.com/articles/google-poaches-aws-microsoft-executives-for-cloud-unit)

🏆Scale AI introduces SEAL Leaderboards, a new evaluation metric for frontier AI models

Trusted third-party evaluations are a missing part of the AI ecosystem, which is why Scale AI built them. These evals will be impossible to overfit, private, domain expert evaluations, unexploitable, and continuously updated with new data and models. (https://x.com/alexandr_wang/status/1795857651592491281)

🎁OpenAI offers nonprofit organizations discounts on corporate ChatGPT subscriptions

It seeks to grow sales of its AI product to enterprises. Under the program, large nonprofits can get 50% off the enterprise-grade version of ChatGPT. Smaller nonprofits using ChatGPT Team will pay $20 per month per user instead of $25 or $30. (https://www.reuters.com/technology/openai-offers-nonprofits-discounts-corporate-chatgpt-product-2024-05-30)

A  Daily chronicle of AI Innovations May 30th 2024:

😟 Microsoft is worried about OpenAI’s deal with Apple

💻 Mistral announces Codestral, a code-generation LLM it says outperforms all others

🚨 FBI dismantles the ‘largest botnet ever’

📰 OpenAI’s news deals continue, with Vox and the Atlantic signing on

🤖 Mistral AI introduces Codestral, a code-generation AI model

📰 OpenAI licenses content from Vox Media and The Atlantic

💻 Google infuses Chromebook Plus with powerful AI features

🤖 Mistral AI introduces Codestral, a code-generation AI model

Codestral is a 22B parameter generative AI model designed specifically for code generation tasks. It is trained in over 80 programming languages, including popular ones like Python, Java, C++, and JavaScript. It excels at code completion, writing tests, filling in partial code, and testing code, thereby improving developer productivity and reducing errors.

The model sets a new performance/latency standard for code generation tasks compared to previous models. It can be downloaded on HuggingFace and is available on Mistral’s API platform through instruct and fill-in-the-middle endpoints. It can be easily integrated into VScode plugins and used for free on Le Chat.

Why does it matter?

Codestral, being only 22B in size and faster than GPT-4o, has massive implications. It is code-centric, runs locally, includes specific languages like Swift and Fortran, and is open for research/testing purposes—all of which make a powerful AI coding tool more widely available.

Source: https://mistral.ai/news/codestral

OpenAI licenses content from Vox Media and The Atlantic 

OpenAI has formed content and product partnerships with Vox Media and The Atlantic. OpenAI will license content from these media powerhouses for inclusion in the chatbot’s responses.

In turn, The Atlantic’s product team will have privileged access to OpenAI tech, give feedback, and share use cases to shape and improve future news experiences in ChatGPT and other OpenAI products. Vox and OpenAI will also collaborate using OpenAI’s technology to develop innovative products for Vox Media’s consumers and advertising partners.

Why does it matter?

There’s a growing list of publishers and platforms that are allowing OpenAI to access its valuable content in mutually beneficial partnerships. It could help OpenAI avoid further legal disputes (like the ongoing lawsuit with The New York Times over copyright infringement) and get quality data to train its LLMs and multimedia models.

Source: https://venturebeat.com/ai/openai-partners-with-the-atlantic-and-the-verge-publisher-vox-media

Google infuses Chromebook Plus with powerful AI features

Google’s Chromebook Plus has new built-in Google AI and gaming features, Help me write, Chat with Gemini, Magic Editor, and more.

  • “Help me write” allows users to get AI-generated text suggestions, tone changes, rewriting, and more right within their writing apps on the Chromebook.
  • Generative AI wallpaper and video call backgrounds are built into the OS for customization, no matter what video conferencing app you’re using.
  • Magic Editor on Google Photos for laptops exclusively on Chromebook Plus will reimagine your photos with a few easy clicks.
  • Chat with Gemini on the home screen helps plan, write, learn, and more.

Chromebook Plus laptops are starting at $350 USD. Google is also launching new features that integrate more of Google across all Chromebooks to help users work better and get things done faster.

Ace the Microsoft Azure Fundamentals AZ-900 Certification Exam: Pass the Azure Fundamentals Exam with Ease

Why does it matter?

It indicates that the next wave of computers will come equipped with advanced AI capabilities. Such widespread availability of AI can improve productivity and enhance user experience while making AI a standard part of everyday life.

Source: https://blog.google/products/chromebooks/chromebook-plus-google

😟 Microsoft is worried about OpenAI’s deal with Apple

    • Microsoft is reportedly worried about Apple collaborating with OpenAI to integrate AI technology into upcoming versions of iOS and macOS, potentially impacting Microsoft’s services and Azure cloud platform.

Microsoft has invested billions in OpenAI, securing an exclusive license for GPT-4 and other models, and uses its Azure cloud as the primary platform for OpenAI’s large language model development.

💻 Mistral announces Codestral, a code-generation LLM it says outperforms all others

  • Mistral introduces Codestral, a 22B parameter AI code assistant trained on more than 80 programming languages, including Swift and Python, designed to help developers write and interact with code effectively.
  • Codestral offers a larger 32K context window and performs exceptionally well in the RepoBench evaluation for long-range code generation, outpacing existing models in various benchmarks and languages.
  • Available for free via Mistral’s conversational AI platform Le Chat and through an API, Codestral is integrated with popular developer tools and licensed under the Mistral AI Non-Production License, restricting its use to research and testing only.
  • Source: https://www.maginative.com/article/mistral-unveils-codestral-an-ai-code-assistant-trained-on-80-programming-languages/

📰 OpenAI’s news deals continue, with Vox and the Atlantic signing on

  • The Atlantic and Vox Media have entered into agreements with OpenAI, allowing the AI company to scrape and use their content in exchange for citations and links back to their original sources.
  • The Atlantic will serve as a “premium news source” for OpenAI, while also launching an experimental section called Atlantic Labs to showcase new journalism-related products and features.
  • Vox Media will integrate OpenAI data into its internal operations and public-facing tools, enhancing content such as the Strategist Gift Scout tool and their in-house advertising platform for better-targeted ads.
  • Source: https://www.engadget.com/the-atlantic-and-vox-media-made-their-own-deal-with-the-ai-devil-161017636.html?

What Else Is Happening in AI on May 30th 2024❗

🚀 SambaNova sets new Llama 3 speed record with 1,000 tokens per second

The feat was accomplished using their SN40L chip, an RDU, and the Samba-1 model, a 1-trillion parameter model also known as Samba-CoE. Such Gen AI performance speed can potentially lead to significant business benefits, such as faster response times, better hardware utilization, and lower costs. (https://venturebeat.com/ai/sambanova-breaks-llama-3-speed-record-with-1000-tokens-per-second)

🛡️ Apple will process data from AI applications in a virtual black box

Apple is set to introduce Apple Chips in Data Centers (ACDC). It involves processing AI data within a virtual black box, preventing employee access to ensure utmost privacy. Apple gains greater control over system design and develops more secure AI platforms by using its custom chipsets for servers and devices(https://in.mashable.com/tech/76158/apple-plans-to-protect-user-data-by-putting-it-into-virtual-black-box-report)

🤝 SAP teams up with Amazon Bedrock for enhanced Gen AI

The collaboration simplifies AI model deployment and ensures compliance with regulatory standards within SAP’s BTP. With features like multitenancy and seamless integration with SAP applications, the integration supports tasks such as model training, inference, and deployment. (https://venturebeat.com/ai/sap-adds-amazon-bedrock-into-ai-core-streamlining-generative-ai-use-for-regulated-firms)

💼 ChatGPT free tier now offers premium features

If you are looking for an all-in-one solution to help you prepare for the AWS Cloud Practitioner Certification Exam, look no further than this AWS Cloud Practitioner CCP CLF-C02 book

Free ChatGPT users can access several advanced features previously reserved for paid subscribers, such as custom GPTs, data analytics, chart creation, vision capabilities, and Memory. However, they can’t create custom GPTs of their own. (https://www.theverge.com/2024/5/29/24167436/chatgpt-4o-custom-gpts-free)

📱 ARM’s new chip designs and software for AI on smartphones

Arm Holdings launched new designs for CPUs and GPUs to help smartphones handle AI tasks better. It will also provide software tools to make it easier for developers to run chatbots and other AI code on Arm chips. https://www.reuters.com/technology/arm-offers-new-designs-software-ai-smartphones-2024-05-29

AI Training: 📊 Unlock ChatGPT’s interactive charts and tables

OpenAI recently introduced a game-changing feature in ChatGPT that lets you analyze, visualize, and interact with your data without the need for complex formulas or coding.

  1. Head over to ChatGPT and select GPT-4o.
  2. Upload data by clicking the 📎 button and select specific columns or rows for a clear focus.
  3. Prompt for interactive charts for powerful visualizations (e.g., “Create a pie chart for X distribution”).
  4. Hover over the different sections to see the exact values, change the chart type, and explore the legend for better understanding

Source: https://university.therundown.ai/c/daily-tutorials/chatgpt-has-now-interactive-charts-and-tables-6d71dd73-80be-4509-97a6-0d16dd1abf04

🧠 AI RESEARCH: AI brain implant language breakthrough

Researchers at UC San Francisco just developed a brain implant that utilizes AI to help a stroke survivor communicate in both Spanish and English, switching between languages seamlessly via brain activity.

  • The bilingual implant was tested on a patient who lost his ability to speak after suffering a stroke at the age of 20.
  • An AI-powered decoding system was trained to recognize the patient’s brain activity patterns when articulating words in both languages.
  • The system determined the patient’s intended language with 88% accuracy and the correct sentence 75% of the time.
  • The implant allows the patient to participate in bilingual conversations and switch between languages, despite not learning English until after his stroke.

This research is another example of AI’s increasing ability to interpret our brainwaves — potentially unlocking an endless supply of new learnings, treatments, and technology. It’s also yet another massive leap for unlocking communication for stroke victims while breaking language barriers in the process.

New AI Job Opportunities on May 30th 2024

A  Daily chronicle of AI Innovations May 29th 2024:

💥 Google faces major search algorithm leak

👀 Former OpenAI board member explains why they fired Sam Altman

🔄 Anthropic recruits ex-OpenAI safety chief to lead new ‘Superalignment’ team

🔒 OpenAI forms an AI safety committee

👀 OpenAI begins training the next model
🔒All ChatGPT Free users can now use browse, vision, data analysis, file uploads, and GPTs.

💥 Google faces major search algorithm leak

  • A leak involving 2,500 pages of internal documentation from Google has surfaced, offering a detailed and unprecedented look into the inner workings of its search algorithm, which is one of the most influential systems on the internet.
  • The leaked documents suggest discrepancies between Google’s public statements and its internal practices, specifically highlighting the use of Chrome data and the importance of author bylines in search rankings, contrary to what Google has previously disclosed.
  • Despite multiple requests for comment, Google has not confirmed the authenticity of the leaked documents, which has led to increased scrutiny and calls for greater transparency.
  • Source: https://www.theverge.com/2024/5/28/24166177/google-search-ranking-algorithm-leak-documents-link-seoGoogle  

👀 Former OpenAI board member explains why they fired Sam Altman

  • Former board member Helen Toner revealed on a podcast that the board lost trust in Altman due to his secret ownership of the OpenAI Startup Fund, providing inaccurate safety information, and personal retaliation against her.
  • Toner explained that after two executives reported a toxic atmosphere and psychological abuse by Altman, the board decided to act, citing the launch of ChatGPT without prior board knowledge, learning about it only through Twitter, as an example of their lack of oversight.
  • She believes pressure to reinstate Altman stemmed from limited options presented to employees, fear of retaliation, and Altman’s troubled history with previous jobs
  • Shttps://www.theverge.com/2024/5/28/24166713/openai-helen-toner-explains-why-sam-altman-was-fired

🔄 Anthropic recruits ex-OpenAI safety chief to lead new ‘Superalignment’ team

  • Jan Leike, a prominent AI researcher who recently resigned from OpenAI, has joined Anthropic to lead a new “superalignment” team focused on AI safety and security.
  • Leike’s team will work on scalable oversight, weak-to-strong generalization, and automated alignment research, reporting directly to Chief Science Officer Jared Kaplan.
  • Anthropic aims to distinguish itself as more safety-focused than OpenAI, with Leike’s new team echoing the mission of OpenAI’s dissolved Superalignment team.
  • Source: https://www.neowin.net/news/former-openai-safety-head-jan-leike-joins-rival-firm-anthropic/

🔒 OpenAI forms an AI safety committee

OpenAI Board formed a Safety and Security Committee led by directors Bret Taylor, Adam D’Angelo, Nicole Seligman, and Sam Altman. Over the next 90 days, the committee will evaluate and further develop OpenAI’s processes and safeguards. Then, the committee will share its recommendations with the full Board.

Following the full Board’s review, OpenAI will publicly share an update on adopted recommendations. OpenAI has recently begun training its next frontier model, and the resulting systems from the Safety and Security team’s recommendation may bring it safely to the next level of capabilities on its path to AGI.

Why does it matter?

OpenAI has drawn a lot of criticism for for putting AI safety on the backseat after its super alignment team was dissolved a few days ago. However, it seems that OpenAI is trying to rise to the occasion and address these concerns while developing AI more responsibly.

Source: https://openai.com/index/openai-board-forms-safety-and-security-committee

👀 OpenAI begins training the next model

OpenAI just announced the formation of a new Safety and Security Committee to oversee the development of its next frontier AI model, which the company also said it has recently started training.

  • The new committee is led by CEO Sam Altman and includes board members Bret Taylor, Adam D’Angelo, and Nicole Seligman.
  • The group will provide evaluation of safety and security processes over the next 90 days, with plans to share adopted recommendations publicly.
  • OpenAI said the new model will ‘bring us to the next level of capabilities on our path to AGI’.

Source: https://openai.com/index/openai-board-forms-safety-and-security-committee

💻 Google Chromebooks get AI infusion

 Google just unveiled a suite of new AI-powered features for its $350 Chromebook Plus laptops, aimed at enhancing productivity, creativity, and collaboration for users.

  • Google’s Gemini assistant is integrated directly into the Chromebook Plus home screen, allowing quick access to AI tools.
  • The “Help Me Write” feature brings AI capabilities to all text entry fields, providing suggestions, changing tone, and rewriting text.
  • Google Photos’ Magic Editor gives users advanced AI-powered image editing capabilities directly within the Chromebook Plus.
  • Generative features like AI wallpapers and video call backgrounds also offer users new customization options.
  • Source: https://blog.google/products/chromebooks/chromebook-plus-google

OpenAIAll ChatGPT Free users can now use browse, vision, data analysis, file uploads, and GPTs.

2 weeks ago

We’re opening up access to our new flagship model, GPT-4o, and features like browse, data analysis, and memory to everyone for free (with limits). Plus users will get up to 5x higher limits, and earliest access to features like our new macOS desktop app and next-generation voice and video capabilities.

Abstract impressionist painting featuring layers of blue and lavender hues, representing a serene seascape.

Introducing GPT-4o and more tools to ChatGPT free users

New AI Job Opportunities May 29th 2024:

  • 🧑‍🎨 Parloa – CX Design Consultant: https://jobs.therundown.ai/jobs/61082222-cx-design-consultant-f-m-d-*
  • 👨‍💻 DeepL – Engineering Manager: https://jobs.therundown.ai/jobs/61074472-engineering-manager-%7C-core-platform-backend
  • ⚖️ Abridge – Assistant General Counsel: https://jobs.therundown.ai/jobs/61108336-assistant-general-counsel
  • 🤝 Writer – Enterprise Business Development Rep: https://jobs.therundown.ai/jobs/61154629-enterprise-business-development-rep

What Else Is Happening in AI on May 29th 2024❗

💼 PwC set to become the first reseller of ChatGPT Enterprise

Companies no longer need to buy a ChatGPT Enterprise license directly from OpenAI. They can purchase the Gen AI service through PwC. Until today, businesses could only subscribe to the enterprise option by contacting an OpenAI salesperson. This also allows PwC to upsell its services to those who want to use ChatGPT to optimize their workloads. (https://venturebeat.com/ai/pwc-strikes-openai-deal-to-become-the-first-reseller-of-chatgpt-enterprise)

🤖 Microsoft brings Copilot AI chatbot to Telegram users

Microsoft has added an official Copilot bot within the messaging app Telegram, which lets users search, ask questions, and converse with the AI chatbot. Copilot for Telegram is currently in beta but is free for Telegram users on mobile or desktop. (https://www.theverge.com/2024/5/28/24166451/telegram-copilot-microsoft-ai-chatbot)

🌐 Opera is integrating Google’s Gemini models into its Aria browser AI

Opera announced a collaboration with Google Cloud to integrate Gemini models into its Aria browser AI. Aria is powered by Opera’s multi-LLM Composer AI engine, allowing it to curate the best user experiences based on their requirements. Thanks to this integration, Opera can now provide its users with the most current information at high performance. (https://press.opera.com/2024/05/28/opera-google-cloud-aria-gemini/)

🚀 GitHub Accelerator program empowers the open-source AI revolution

GitHub launched the 2024 Accelerator program which provides funding, mentorship, and community-building. Standout participants include Unsloth, which improves AI model efficiency, and Formbricks, which simplifies user feedback. Through this, GitHub accelerates open-source innovation and democratizes access to new tech. (https://venturebeat.com/ai/github-accelerator-fuels-open-source-ai-revolution-empowering-startups-to-democratize-access)

🤼‍♂️ Elon Musk vs. Yaan LeCun on X highlights differences in AI research approach

Elon Musk and Yann LeCun, two prominent figures in AI, got into a heated debate on X over the weekend. LeCun criticized Musk’s management style at his new AI startup xAI, while Musk questioned LeCun’s recent scientific contributions. LeCun emphasizes the importance of open scientific publication, while Musk focuses on ambitious goals like artificial general intelligence (AGI). (https://venturebeat.com/ai/elon-musk-and-yann-lecuns-social-media-feud-highlights-key-differences-in-approach-to-ai-research-and-hype/))

 A  Daily chronicle of AI Innovations May 28th 2024:

🚨 Google AI Overview gives dangerous and wrong answers
🚀 GPT-4 is a better financial analysts than humans
🎨 Canva redesigned it’s platform with new AI features

🇪🇺 The EU’s data protection task force has released preliminary findings on ChatGPT’s GDPR compliance 

📱 French startup AniML has launched Doly, an iPhone app simplifying 3D product video creation

🎧 Iyo, an Alphabet X spinout, will release the Iyo One, a pair of gen AI earbuds 

💳 Mastercard integrates AI to catch compromised cards faster

🧮 Peter Thiel, former PayPal CEO, believes AI will worse for math minds more than  writers

🚨 Google AI Overview gives dangerous and wrong answers

Google’s AI Overviews feature, which generates AI-powered responses to user queries, has been providing incorrect and sometimes bizarre answers. From suggesting glue on pizza, staring at sun for health benefits to claiming that former US President Barack Obama is Muslim, the feature has left users questioning the reliability of AI-generated search results.

While Google maintains that these mistakes result from uncommon queries and are being used to refine the product, the technology’s widespread deployment has highlighted the challenges and risks of integrating AI into search engines.

Why does this matter?

Other companies, such as OpenAI, Meta, and Perplexity, have also experienced issues with AI hallucinations and mistakes. Companies must prioritize the development of robust safeguards and rigorous testing to ensure that AI-powered search results meet the high standards users expect from traditional search engines.

Source: https://techcrunch.com/2024/05/26/what-are-googles-ai-overviews-good-for/

🚀 GPT-4 is a better financial analysts than humans

In a groundbreaking study, researchers at the University of Chicago have discovered that GPT-4, a large language model, can analyze financial statements and predict future earnings direction with remarkable accuracy. Even without narrative context, GPT-4 outperforms human financial analysts and achieves prediction accuracy on par with state-of-the-art machine learning models.

The AI model’s success is not simply a result of its training memory but rather its ability to generate valuable narrative insights about a company’s future performance. Notably, trading strategies based on GPT-4’s predictions yield higher returns and risk-adjusted performance than those based on other models, especially for small companies.

Why does this matter?

This study shows that AI is more than just a support tool in financial decision-making. It can also be central to financial analysis and enable non-experts to make informed decisions, which may change how financial markets operate.

Source: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4835311

🎨 Canva redesigned it’s platform with new AI features

Canva is launching a redesigned platform with new AI features and tools for professional teams and workspaces. The company has introduced Canva Enterprise, a specialized tier offering more control over collaboration, brand management, and security for larger organizations.

Ad Tech integrations with Google, Meta, and Amazon streamline the ad creation process within Canva. Data autofill automates the creation of data-driven designs by integrating with sources like Salesforce and MLS.

New features include AI style matching for brands, customizable folder displays, and the ability to “star” designs and templates for easier access. Canva Docs now has a suggestion mode for editors and colorful highlight blocks for text emphasis. New “Magic Studio” AI tools include:

  • Automatic clip highlighting.
  • Background noise reduction for video editing.
  • A text-to-graphic image generator.

Why does it matter? 

Canva aims to reduce organizational complexity and “app sprawl” by offering a comprehensive creative platform. This update addresses the growing need for extensive, user-friendly solutions that can replace multiple design, AI, and workflow apps.

Source: https://www.canva.com/newsroom/news/canva-for-work

What Else Is Happening in AI on May 28th 2024❗

🇪🇺 The EU’s data protection task force has released preliminary findings on ChatGPT’s GDPR compliance

While the group remains undecided on key legal issues, it suggested that “adequate safeguards” and “precise collection criteria” could help OpenAI meet the requirements. However, the AI giant’s move to Ireland may benefit from the country’s business-friendly approach to GDPR enforcement. (Link: https://techcrunch.com/2024/05/27/eus-chatgpt-taskforce-offers-first-look-at-detangling-the-ai-chatbots-privacy-compliance)

📱 French startup AniML has launched Doly, an iPhone app simplifying 3D product video creation

Doly uses AI to estimate the 3D shape of an object from regular 2D photos, and then it applies a technique called Gaussian splatting to create a high-quality 3D model suitable for use in product videos. Users can capture a 3D model, choose a template from the library, and integrate their object into a 3D scene. (Link: https://techcrunch.com/2024/05/27/doly-lets-you-generate-3d-product-videos-from-your-phone)

🎧 Iyo, an Alphabet X spinout, will release the Iyo One, a pair of gen AI earbuds 

The Iyo One will be released this winter, starting at $599, and aims to succeed, whereas competitors like Humane’s Ai Pin and Rabbit’s R1 have struggled. It integrates LLM-based models for a more sophisticated AI experience. Iyo’s founder, Jason Rugolo, believes the Iyo One will provide value through its sound isolation, comfort, and music quality, in addition to its AI features.(Link: https://techcrunch.com/2024/05/27/iyo-thinks-its-gen-ai-earbuds-can-succeed-where-humane-and-rabbit-stumbled)

💳 Mastercard integrates AI to catch compromised cards faster

Mastercard is rolling out an AI update to its fraud-prediction technology to identify compromised credit and debit cards before criminals use them. By analyzing patterns and contextual data, the AI can proactively flag stolen cards, allowing banks to replace them and protect customers from fraudulent transactions. (https://apnews.com/article/mastercard-visa-ai-credit-card-fraud-detection-0c348818087a57b13bfac66c761e03b4)

🧮 Peter Thiel, former PayPal CEO, believes AI will worse for math minds more than  writers

Thiel states that Silicon Valley in the 21st century is too biased towards math people. Thiel  explained that the exact reason for ‘AI being not able to overtake human creativity,’ is not clear bets on getting worse.”He concludes that AI is not yet good at math, although its capabilities are increasing, as shown by ChatGPT scoring 96% in a UK A-level math paper with the Wolfram plug-in. (Link: https://www.financialexpress.com/life/technology-former-paypal-ceo-says-ai-may-be-good-at-maths-but-it-cant-be-a-great-writernbsp-3503476)

AI researcher Kai-Fu Lee doubled down on his 2017 prediction that AI would displace 50% of jobs by 2027, saying white collar jobs will be eliminated faster than blue collar work.

New AI Job Opportunities May 28th 2024

A  Daily chronicle of AI Innovations May 27th 2024:

💥 Elon Musk’s xAI raises $6B to build ‘Gigafactory of Compute’

🔮 Apple bets that its giant user base will help it win in AI

💰 China invests $47 billion in largest ever chip fund

📊 GPT-4 surpasses humans in financial analysis

🕵️ Microsoft’s Recall AI: Efficiency upgrade or privacy nightmare?

🎧 AI-powered headphones let you listen to one person in a crowd

⚡ xAI’s supercomputer that’s 4x larger than existing GPU clusters

💥 Elon Musk’s xAI raises $6B to build ‘Gigafactory of Compute’

  • Elon Musk’s xAI has successfully raised $6 billion in a Series B funding round to construct a supercomputer known as the “Gigafactory of Compute,” which will be powered by 100,000 Nvidia H100 GPUs, making it at least four times larger than the largest existing GPU clusters.
  • This funding will enable xAI to advance its product offerings, develop cutting-edge infrastructure, and accelerate research and development, with investors including Andreessen Horowitz, Sequoia Capital, and Saudi Prince Alwaleed bin Talal.
  • The supercomputer will support the next iteration of xAI’s chatbot, as xAI aims to create advanced AI systems that are truthful, competent, and maximally beneficial for humanity, continuing Musk’s vision of a “maximum truth-seeking AI” called TruthGPT.
  • Source

🔮 Apple bets that its giant user base will help it win in AI

  • Apple is betting on its vast user base to give it an edge in the AI market, despite its first set of AI features not being as advanced as those from other competitors like Microsoft, Google, and OpenAI.
  • The company plans to introduce AI tools integrated into its core apps and operating systems, focusing on practical, everyday uses for consumers, with much of the AI processing done on-device and more intensive tasks handled via the cloud.
  • Apple’s collaboration with OpenAI and potential agreements with Google indicate it is relying on partnerships to compete in the AI space while its own AI developments are still maturing, leveraging its extensive user base to rapidly scale the use of new AI features.
  • Source

💰 China invests $47 billion in largest ever chip fund

  • China has invested more than $47 billion into its largest-ever chip investment fund to pursue self-sufficiency in semiconductor manufacturing.
  • This significant investment reflects China’s broader strategy to develop homegrown chipmakers amid rising technological competition with the United States.
  • The fund’s third phase, supported by China’s finance ministry and state-owned banks, marks the largest investment of its kind, surpassing the combined total of the previous two phases.
  • Source

📊 GPT-4 surpasses humans in financial analysis

  • New research from the University of Chicago shows that GPT-4 can analyze financial statements more accurately than humans, achieving a 60% accuracy rate compared to human analysts’ 53-57% range.
  • The study indicates significant implications for the future of financial analysis, highlighting GPT-4’s versatility in performing tasks usually reserved for specialized tools and stating its performance is on par with leading machine learning models.
  • While GPT-4 excels in quantitative analysis, human analysts provide valuable contextual insights, suggesting a complementary relationship between AI and humans for achieving optimal financial analysis results.
  • Source

Microsoft’s Recall AI: Efficiency upgrade or privacy nightmare?

Microsoft has recently released Recall, a controversial AI-powered tool that logs everything you see and do on your computer, including app usage, live meeting communications, websites visited, and more. By simply performing a “Recall” action, users can retrieve any information they’ve interacted with on their device, presented in the context of a specific time period.

Microsoft assures users that the Recall index remains local and private on-device. Users can pause, stop, or delete captured content and can choose to exclude specific apps or websites. And Recall automatically excludes InPrivate web browsing sessions in Microsoft Edge and DRM-protected content.

Currently, Recall is exclusively compatible with new “Copilot Plus PCs” equipped with Qualcomm’s Snapdragon X Elite chips, which feature the necessary neural processing unit (NPU).

Why does this matter?

The Recall feature has sparked concern among cybersecurity experts and government authorities due to its extensive data collection and privacy risks. Recall continuously captures screenshots of users’ PC activities and creates a searchable index using AI, which may accidentally expose sensitive information to unauthorized users.

Source

AI-powered headphones let you listen to one person in a crowd

Researchers at the University of Washington have developed a groundbreaking artificial intelligence system called “Target Speech Hearing” that allows headphone wearers to isolate and listen to a single speaker in a noisy environment. By simply looking at the desired speaker for three to five seconds, the user can “enroll” them, and the system will cancel out all other sounds, playing only the enrolled speaker’s voice in real-time, even as the listener moves around.

The proof-of-concept device, which uses off-the-shelf headphones fitted with microphones and an on-board embedded computer, builds upon the team’s previous “semantic hearing” research. The system’s ability to focus on the enrolled voice improves as the speaker continues talking, providing more training data. While currently limited to enrolling one speaker at a time and requiring a clear line of sight, the researchers are working to expand the system to earbuds and hearing aids in the future.

Why does this matter?

The Target Speech Hearing system may have applications in various settings, like business meetings, conferences, noisy public spaces, or for people with hearing difficulties. While the current system is a proof-of-concept, the researchers’ plan to expand it to earbuds and hearing aids indicates the potential for commercialization, which could lead to a new-age consumer product that enhances our sensory experiences in previously unimaginable ways.

Source

xAI’s supercomputer that’s 4x larger than existing GPU clusters

Elon Musk has recently disclosed plans to build a groundbreaking supercomputer to power the next iteration of xAI’s chatbot, Grok. According to a presentation made to investors in May, as reported by The Information, Musk aims to have this computing system operational by the fall of 2025. The ambitious project may involve a collaboration between xAI and Oracle to develop this massive computing infrastructure.

The proposed supercomputer will consist of interconnected clusters of Nvidia’s state-of-the-art H100 graphics processing units (GPUs). Musk revealed that upon completion, the scale of this system will surpass the largest existing GPU clusters by at least fourfold!

Why does it matter? 

As major tech giants and startups compete for dominance in the AI space, Musk’s proposed “gigafactory of compute” could potentially set new benchmarks for AI computing power. Also, Musk’s financial resources and personal commitment to the project position him as a formidable challenger to established AI powerhouses like OpenAI, Google, and Meta.

Source

New AI Jobs Opportunities on May 27th 2024

  • 🤿 Snorkel – Machine Learning Customer Engineer
  • 🐙 OctoAI – Senior MLSys Engineer
  • 🌐 OpenAI – Technical Program Manager, Trustworthy AI
  • 📊 C3 AI – Senior Revenue Accountant

What Else Is Happening in AI on May 27th 2024❗

🎵 YouTube Music introduces hum-to-search AI feature

YouTube Music has introduced a new AI feature that lets users search for songs by humming or singing a melody. This feature, similar to the “Hum to Search” functionality in Google Search, uses machine learning to match the user’s audio input to the original song recording. The feature is currently available in its Android app version 7.02 and is being gradually rolled out to users. (Link)

🤩 iOS 18 may introduce AI-powered custom emoji tool and smart recaps

Apple is reportedly working on a gen AI-powered custom emoji tool for iOS 18 that lets users create personalized emojis. They’re also developing a smart recaps feature that will provide users with summaries of missed notifications, messages, and documents. (Link)

🎼 Suno.ai’s music generator now offers 4-minute songs

AI music model Suno.ai has released its version 3.5 which lets users create songs up to four minutes long with improved structure. The update also introduces a sound-to-song feature, combining audio and text prompts to generate music. (Link)

🔍 Meta to use EU user data for AI training, offers opt-out

Meta is notifying Facebook and Instagram users in Europe about changes to its privacy policy, effective June 26, which allows the company to use user data for training AI models under a “legitimate interest” claim. Users can opt out of this data usage, but they must provide an explanation of how it personally affects them, with a reference to the GDPR being sufficient. (Link)

📱 Google launches Gemini AI assistant in Messages app

Google has started rolling out its Gemini AI assistant feature in Google Messages to help users with tasks like drafting messages, brainstorming, event planning, and engaging in conversations. Gemini supports extensions like Workspace, YouTube, and Google Maps, and also lets users provide feedback on responses. (Link)

A  Daily chronicle of AI Innovations May 24th 2024: 🍕 Google AI tells users to glue pizza and eat rocks ✌️ OpenAI scraps controversial nondisparagement agreement with employees 💰 Musk’s xAI nears deal valuing startup at $24B🌍 Cohere releases multilingual AI model, Aya 23 📱 Arc introduces “Call Arc” for quick voice answers 🤖 Elon Musk envisions AI era, new work norms, life on Mars 🤯 OpenAI controversy: ‘Sky’ speaks out

🍕 Google AI tells users to glue pizza and eat rocks

  • Google’s new AI Overviews feature is generating incorrect and sometimes absurd answers, like adding non-toxic glue to pizza as a solution to prevent cheese from falling off.
  • The feature has made several other mistakes, such as stating that former US President James Madison graduated from the University of Wisconsin 21 times and claiming that Batman is a cop.
  • Google acknowledges these are “isolated examples” and emphasizes the AI is experimental, but the errors highlight significant issues with the current state of AI-generated information.
  • Source

✌️ OpenAI scraps controversial nondisparagement agreement with employees

  • OpenAI will not enforce any previously signed nondisparagement agreements with former employees and will remove such language from its exit paperwork, as stated to Bloomberg.
  • Previously, employees had to choose between speaking against the company or keeping their vested equity, potentially losing millions if they refused to sign the agreement.
  • Sam Altman, OpenAI’s CEO, expressed embarrassment over the existence of the clause and pledged to rectify the paperwork, while Chief Strategy Officer Jason Kwon apologized for the distress caused.
  • Source

💰 Musk’s xAI nears deal valuing startup at $24B

  • Elon Musk’s AI startup, xAI, is expected to complete a funding round in June that could value the company at over $24 billion.
  • The company originally aimed to raise around $6 billion earlier this month but is now seeking to gather as much as $6.5 billion, with final deal closure still pending.
  • xAI has been marketed to Silicon Valley investors using a pitch highlighting Musk’s success with Tesla and SpaceX, and plans to leverage data from his social platform X (formerly Twitter) to build its AI chatbot, Grok.
  • Source

Cohere releases multilingual AI model, Aya 23

Cohere for AI (C4AI), the non-profit research group, has launched open-weight Aya 23, a new family of multilingual language models. Available in 8B and 35B parameter variants, Aya 23 supports 23 languages, including Arabic, Chinese, English, French, German, Hindi, Japanese, Spanish, and more.

Here’s a quick breakdown: 

  • Aya23 focuses on depth over breadth, meaning it performs better in fewer languages than their previous model, Aya 101 (which covered 101 languages).
  • The 8B parameter model balances efficiency and accessibility, while an advanced 35B parameter delivers higher performance at the cost of increased computational demand.
  • Aya 23 outperforms existing models like Google’s Gemma on various tasks across the languages it covers.
  • Researchers can access and fine-tune Aya 23 for their needs, with the model available for free trial on Cohere Playground.

Why does this matter?

Most AI models struggle with languages besides English. The development of multilingual models can serve a much wider audience, bridging the language divide and making AI more accessible and inclusive for users around the world.

Source

Arc introduces “Call Arc” for quick voice answers

Arc Search, an AI-powered search app, just launched a new feature called Call Arc. This lets users ask questions by holding their phone to their ear, mimicking a phone call. It provides instant voice answers, similar to voice search, but designed to be more convenient and quicker.

The app is designed to answer short, immediate questions. For example, you can ask it how long it takes to cook spaghetti or why to reserve pasta water, all while making dinner.

Call Arc complements Arc Search’s existing “Browse for me” function that generates webpages with information based on your search query.

Why does this matter?

As AI continues to simplify our lives, innovative features like Call Arc could provide a new approach to voice search, making it intuitive and accessible for users on the go. This could also lead to more nuanced search results and a more engaging experience compared to simple keyword searches.

Source

Elon Musk envisions AI era, new work norms, life on Mars

In a Q&A session at VivaTech 2024, Elon Musk discussed diverse topics, from plans for Mars colonization to the role of AI in society.

Source: Viva Technology

Musk emphasized SpaceX’s goal of making life multi-planetary, with Mars as a key focus. He discussed the importance of reusable spacecraft and highlighted the necessity of space exploration for humanity’s long-term survival.

Regarding AI, Musk stressed the importance of honesty in AI development, criticizing approaches that prioritize political correctness over truthfulness. He also touched on AI’s potential to revolutionize education, though he expressed concerns about the impact of social media on children.

Musk envisioned a future where automation leads to a job-free society, with a universal basic income ensuring people’s needs are met.

Why does this matter?

Elon Musk’s bold predictions about Mars, the Moon, and a job-free future driven by AI are both exciting and thought-provoking. While his vision for Mars could redefine human civilization, his stance on AI emphasizes the need for integrity in technology.

It’s crucial that we engage in open and honest dialogue about AI’s implications for society.

Source

🤯 OpenAI controversy: ‘Sky’ speaks out

There has been a wave of updates on several drama-filled OpenAI topics, including leaked documents related to the company’s controversial NDA clauses, new info surrounding the ‘Sky’ voice model, another resignation, and more.

  • The voice actress behind ChatGPT’s ‘Sky’ voice has come forward, saying she was hired months before Altman contacted Scarlett Johansson.
  • Leaked documents with Sam Altman’s signature contradict claims he was unaware of NDA clauses threatening to claw back former employees’ equity.
  • OpenAI policy researcher Gretchen Kreuger resigned, with a post citing safety concerns about the company.
  • Teasers are coming out of presentations at the VivaTech conference, showing a potential move away from the ‘GPT-5’ name and a new Sora demo.

New AI Jobs Opportunities on May 24th 2024:

  • 🌐 Kumo – ML Solutions Architect
  • 📝 Palantir – Proposal Writer, Internship
  • 💻 Lambda – Linux Support Engineer
  • 🏢 Anthropic – Enterprise Account Executive

What Else Is Happening in AI on May 24th 2024❗

🎵 YouTube launches AI-powered music creation tool

YouTube has introduced Dream Track, an AI-powered tool designed for creating music within YouTube Shorts. This experimental feature allows users to generate musical pieces based on input commands. While it is an innovative addition, it has sparked a debate over AI’s impact on the authenticity of creative work.  (Link)

🎨 Ideogram launches Ideogram Tile, an AI-powered pattern generator

Ideogram has launched Ideogram Tile, a feature that creates seamless, repeating patterns from text prompts. This tool empowers designers to craft unique wallpapers, fabrics, and textures effortlessly. Following a $80 million funding round, Ideogram continues to lead in AI-driven design innovation, blending creativity with advanced technology. (Link)

🎧 Spotify tests Spanish-speaking AI DJ, “DJ Livi”

Spotify is testing a Spanish-speaking iteration of its AI DJ, named “DJ Livi.” The app code suggests a language switch option for this new feature. Although Spotify has not confirmed launch plans, this move could cater to Spanish-speaking audiences globally and in the U.S., marking a significant expansion in AI language support. (Link)

🗣️ Amazon plans paid subscriptions for Alexa AI features

Amazon is reportedly gearing up to launch paid subscription plans for select Alexa AI features, aiming to enhance conversational capabilities. This move seeks to compete with advanced AI chatbots like Google’s Gemini and OpenAI’s ChatGPT. Notably, the subscription won’t be bundled with Amazon Prime, offering a standalone service. (Link)

🎮 Inworld AI launches AI-driven voice generator

Inworld AI, known for its innovative AI-driven virtual characters, unveils Inworld Voice, a dynamic voice generator for game developers. Boasting 58 diverse voices powered by advanced machine learning models, it promises expressive, cost-efficient, and high-quality voice experiences. (Link)

A  Daily chronicle of AI Innovations May 23rd 2024: 🔍 Anthropic uncovers millions of concepts in Claude Sonnet’s AI model 📞 Truecaller’s AI assistant gets a voice upgrade, thanks to Microsoft 🎥 TikTok makes ad creation easy with AI! 🤯 Scientists plan a head transplant system 🦎 Chameleon: Meta’s new multimodal LLM 🤔 OpenAI didn’t copy Scarlett Johansson’s voice, records show 📈 Nvidia shows no signs of AI slowdown 💥 Global outages hit Microsoft Bing, DuckDuckGo, ChatGPT Search and Copilot 💰 OpenAI signs $250M landmark deal with News Corp ⛑️ Apple may develop foldable iPhone with self-healing screen 

Anthropic uncovers millions of concepts in Claude Sonnet’s AI model

Anthropic has made a breakthrough in understanding the inner workings of their AI model, Claude Sonnet, by identifying how millions of concepts are represented within it.

Using a technique called “dictionary learning,” they were able to map out these concepts, providing the first-ever detailed look inside a modern, production-grade large language model.

Key findings:

  • Features linked to concepts: These concepts are linked to features, which are groups of neurons that activate together in response to specific ideas.
  • Features can be manipulated: By manipulating these features, the researchers were able to influence Claude’s outputs, demonstrating a causal link between features and behavior.
  • Features reveal potential risks: The research identified features corresponding to biases, potential misuse of the model, and even sycophantic behavior.

Why does this matter?

Anthropic’s research is a big step toward making AI models more transparent and trustworthy. By understanding how these models work, researchers can spot and reduce harmful biases, prevent misuse, and steer AI systems toward better outcomes, leading to safer and more reliable AI applications.

Source

Truecaller’s AI assistant gets a voice upgrade, thanks to Microsoft

Truecaller is partnering with Microsoft to let users create an AI version of their own voice for their AI Assistant to use when answering calls.

This feature, currently limited to users with access to Truecaller’s AI Assistant, requires them to record a voice clip that Microsoft’s Azure AI Speech technology will use to create a personalized AI voice.

Truecaller says this will allow for a more personalized experience and highlights the potential of AI in communication. However, it’s important to note that Microsoft limits the use of personal voice for specific purposes and requires users to obtain consent before recording someone’s voice.

Why does this matter?

Truecaller’s AI voice assistant offers a glimpse into the future of call management, promising a more personalized experience. However, it must ensure transparency and address performance comparisons with automated voicemail systems.

The idea of speaking with an AI while calling someone may seem unusual, and it remains to be seen how well this feature will be received.

Source

TikTok makes ad creation easy with gen AI!

TikTok has introduced “TikTok Symphony,” a suite of generative AI tools designed to help marketers create and optimize ad campaigns. The suite includes an AI video generator called “Symphony Creative Studio,” which can produce TikTok-ready videos with minimal input from advertisers, and an AI assistant named “Symphony Assistant” that helps refine scripts and provides best practice recommendations.

TikTok makes ad creation easy with gen AI!
TikTok makes ad creation easy with gen AI!

The company has also introduced “TikTok One,” a centralized hub for marketers to access creators, agency partners, and creative tools. Additionally, TikTok is leveraging predictive AI to drive more sales for advertisers by determining the best creative assets and target audiences based on budgets and goals.

Why does this matter?

TikTok’s integration of generative AI into its ads business signifies the growing importance of AI in digital marketing. This shift could make ad campaigns more efficient and effective, opening new opportunities for tech companies and startups.

Source

 Chameleon: Meta’s new multimodal LLM

Chameleon: Meta’s new multimodal LLM
Chameleon: Meta’s new multimodal LLM

Meta’s AI research lab just introduced Chameleon, a new family of ‘early-fusion token-based’ AI models that can both understand and generate text and images in any order.

  • Unlike other models that process image and text separately and then combine them later, Chameleon works with linked sequences of both.
  • Chameleon outperformed all rival models on image captioning and visual questions, while still matching performance on text-only.
  • The 34B parameter Chameleon also matched or outperformed top models like Gemini Pro and GPT-4V on tests of long-form mixed-modal generations.

Chameleon shows the potential for a different type of architecture for multimodal AI models, with its early-fusion approach enabling more seamless reasoning and generation across modalities and setting new performance bars.

New AI Jobs Opportunity on May 23rd 2024

  • 🎯 Impel AI – Test Automation Engineer
  • 📚 Grammarly – Researcher, Strategic Research
  • ⛅ Tempus – Senior Cloud Security Engineer
  • 📋 UiPath – Senior Product Manager

What Else Is Happening in AI on May 23rd 2024❗

🤝 OpenAI partners with News Corp to enhance ChatGPT 

OpenAI has announced a multi-year partnership with media giant News Corp to enhance ChatGPT with its high-quality journalism. The collaboration aims to provide reliable information to users, with News Corp’s mastheads displayed in response to user queries. OpenAI CEO Sam Altman called it a “proud moment for journalism and technology.” (Link)

💻 AWS and Hugging Face team up on AI model development

AWS has partnered with AI startup Hugging Face to enable developers to run thousands of AI models on Amazon’s custom Inferentia2 chips. The collaboration aims to provide a cost-effective solution for deploying AI models, with AWS aiming to compete in the inference market currently dominated by Nvidia. (Link)

📝 Granola’s introduces AI-powered note-taking app

Granola, a new AI startup, has launched a notepad app that combines user-written notes with AI-generated details from meeting transcripts. The app aims to make note-taking more collaborative and efficient, using OpenAI’s GPT-4 to augment and clean up your notes. (Link)

🧠 BrainBridge launches head transplant system using AI and robotics

Neuroscience startup BrainBridge has revealed its groundbreaking head transplant system that aims to provide hope for patients with untreatable conditions.

https://x.com/i/status/1793045981954523464

  • The CGI demo proposes a system that uses advanced robotics, AI, and real-time molecular imaging to transfer a patient’s head or face onto a donor body.
  • The company hopes to conduct the first surgery within eight years, and is releasing the concept early to attract top scientists.
  • A headband with a brain-computer interface (like Neuralink) would help patients communicate and execute tasks during recovery using their minds.

While the mind-blowing (🥁) system seems highly speculative and more sci-fi than reality — so did Neuralink’s brain chips. While we’re likely a long way from this becoming possible, with the rate of advancement in robotics and AI it’s tough to rule anything out over the next decade.

The procedure uses advanced AI, robotics, and a proprietary adhesive to transplant a patient’s head onto a donor body. While it’s still early, BrainBridge expects to conduct the first surgery within eight years, pending successful feasibility studies. (Link)

👨‍💻 Meta establishes AI advisory group

Meta CEO Mark Zuckerberg has created the Meta Advisory Group, consisting of Stripe’s Patrick Collison, Shopify’s Tobi Lütke, and others, to guide the company’s AI and technology advancements. The group will offer insights and recommendations as Meta pushes harder into AI development across various products. (Link)

A  Daily chronicle of AI Innovations May 22nd 2024: 🧠 Microsoft’s first SoTA SLM to be shipped with Windows
📈 Google unveils new AI tools for branding and product marketing 🎨 Adobe introduces Firefly AI-powered Generative Remove to Lightroom

Microsoft’s first SoTA SLM to be shipped with Windows

Microsoft announced a new small language model called Phi Silica. It has 3.3 billion parameters, which makes it the smallest model in Microsoft’s Phi family of models. Phi Silica is designed specifically for the Neural Processing Units (NPUs) in Microsoft’s new Copilot+ PCs. Despite its small size, Phi Silica can generate 650 tokens per second using only 1.5 Watts of power. This allows the PC’s main processors to be free for other tasks.

Microsoft's first SoTA SLM to be shipped with Windows
Microsoft’s first SoTA SLM to be shipped with Windows

Developers can access Phi Silica through the Windows App SDK and other AI-powered features like OCR, Studio Effects, Live Captions, and Recall User Activity APIs. Microsoft plans to release additional APIs, including Vector Embedding, RAG API, and Text Summarization. These AI-powered PCs will have dedicated AI chips for running LLMs and other AI workloads.

Why does it matter?

As Microsoft continues to invest in developing small language models and integrating AI into its Windows platform, Phi Silica represents a significant step forward in making advanced AI capabilities more accessible to developers and end-users. Also, as major PC manufacturers want to introduce AI-powered laptops this summer, Microsoft might lead the way by introducing CoPilot PC+ and now Phi Silica.

Source

Google unveils new AI tools for branding and product marketing

Google has introduced several new AI-powered features to help retailers and brands better connect with shoppers. First, Google has created a new visual brand profile that will appear in Google Search results. This profile uses information from Google Merchant Center and Google’s Shopping Graph to showcase a brand’s identity, products, and offerings.

Google unveils new AI tools for branding and product marketing
Google unveils new AI tools for branding and product marketing

Additionally, Google is expanding its AI-powered tools to help brands create more engaging content and ads. This includes new features in Google’s Product Studio, allowing brands to generate images matching their unique style.

Google unveils new AI tools for branding and product marketing
Google unveils new AI tools for branding and product marketing

Google is also launching immersive ad formats powered by generative AI, such as the ability to include short product videos, virtual try-on experiences, and 3D product views directly in search ads. These new AI-driven tools aim to help brands forge stronger, more personalized connections with consumers throughout the shopping journey.

Google unveils new AI tools for branding and product marketing
Google unveils new AI tools for branding and product marketing

Why does it matter?

As AI continues advancing, such tools could shape the future of marketing by delivering more personalized, hyper-relevant, and visually compelling experiences that better connect consumers with brands and products.

Source

Adobe introduces Firefly AI-powered Generative Remove to Lightroom

Adobe has added a new AI-powered feature called Generative Remove to its Lightroom photo editing software. Generative Remove uses Adobe’s Firefly generative AI model to allow users to seamlessly remove objects from photos, even if the objects have complex backgrounds. The feature can remove images’ stains, wrinkles, reflections, and more.

Adobe introduces Firefly AI-powered Generative Remove to Lightroom
Adobe introduces Firefly AI-powered Generative Remove to Lightroom

Adobe has been integrating Firefly’s capabilities across its Creative Cloud apps to generate images, apply styles, fill areas, and remove objects through the new Generative Remove tool in Lightroom. It works closely with photographers to continue improving and expanding this object-removal capability. The company also announced a new Lens Blur effect that uses AI to add realistic depth-of-field blur to photos.

Why does it matter?

The Generative Remove feature will make it easier for photographers, designers, and other creatives to edit their images, saving time and effort. Looking ahead, we can expect Adobe and other creative software companies to explore new ways to harness Gen AI to automate tedious tasks, provide intelligent assistance, and enable entirely new creative possibilities.

Source

AI RESEARCH:

🗺️ Research: Mapping the ‘mind’ of an LLM

Research: Mapping the ‘mind’ of an LLM
Research: Mapping the ‘mind’ of an LLM

Anthropic just published new research that successfully identified and mapped millions of human-interpretable concepts, called “features”, within the neural networks of Claude.

The details:
  • Researchers used a technique called ‘dictionary learning’ to isolate patterns that corresponded to concepts, from objects to abstract ideas.
  • By tweaking the patterns, the researchers showed the ability to change Claude’s outputs, potentially leading to more controllable systems.
  • The team mapped concepts related to AI safety concerns, like deception and power-seeking — providing glimpses into how models understand these issues.

Why it matters: Despite how fast AI is accelerating, we still don’t have a strong understanding of what’s going on beneath the hood of LLMs. This research is a major step towards making AI more transparent — enabling better understanding, control, and safeguarding of these powerful tools.

New AI Job Opportunities on May 22nd 2024

What Else Is Happening in AI on May 22nd 2024❗

🤖 Elon Musk’s xAI plans to make Grok multimodal   

According to public developer documents, Elon Musk’s AI company, xAI, is making progress on adding multimodal inputs to the Grok chatbot. This means soon, users may be able to upload photos to Grok and receive text-based answers. This was first teased in a blog post last month from xAI, which said Grok-1.5V will offer “multimodal models in a number of domains.” (Link)

👨‍💻 Microsoft’s new Copilot AI agents to work like virtual employees 

Microsoft will soon allow businesses and developers to build AI-powered Copilots that can work like virtual employees and perform tasks automatically. Instead of Copilot sitting idle waiting for queries, it will be able to monitor email inboxes and automate tasks or data entry that employees normally have to do manually. (Link)

🌍 Microsoft Edge introduces real-time AI translation and dubbing for YouTube

Microsoft Edge is set to introduce real-time translation and dubbing for videos on platforms like YouTube, LinkedIn, and Coursera. This new AI-powered feature will translate spoken content live, offering dubbing and subtitles. Currently, the feature supports translations from Spanish to English and English to German, Hindi, Italian, Russian, and Spanish. (Link)

🛡️ WitnessAI builds guardrails for Gen AI models

WitnessAI is developing tools to make Gen AI models safer for businesses. The company’s platform monitors employee interactions and custom AI models, applying policies to reduce risks like data leaks and biased outputs. The platform also offers modules to enforce usage rules and protect sensitive information. Lastly, it encrypts and isolates data for each customer. (Link)

💻 Microsoft’s Azure AI Studio supports GPT-4o

Microsoft has announced that Azure AI Studio is now generally available and supports OpenAI’s GPT-4o model, which joins over 1,600 other models, including Mistral, Meta, Nvidia, etc. Developers can use this multimodal foundational model to incorporate text, image, and audio processing into their apps to provide generative and conversational AI experiences. (Link)

  Daily chronicle of AI Innovations May 21st 2024: 💥 Microsoft announces AI-powered MacBook competitors 🧠 Microsoft’s AI chatbot will ‘recall’ everything you do on a PC 🫠 Scarlett Johansson told OpenAI not to use her voice 🔧 TSMC and ASML machines are equipped with remote self-destruct in case of invasion 🚚 Volvo introduces first autonomous truck 💻 Microsoft’s New AI PCs Rival Apple’s MacBooks ⚖️ Scarlett Johansson sues OpenAI for using her voice in ChatGPT 🧠 DINO 1.5 is smarter and faster at object detection

💥 Microsoft announces AI-powered MacBook competitors

  • Microsoft introduces “Copilot Plus PCs,” new Windows laptops with built-in AI hardware and features, launching June 18th with Qualcomm processors.
  • These AI-enabled laptops, from major partners like Dell and Lenovo, will include a neural processor, at least 16GB RAM, and 256GB SSD, supporting over 40 AI models and OpenAI’s GPT-4o model.
  • Promising significant performance and battery life improvements, Microsoft expects to sell 50 million Copilot Plus PCs in the next year, marking a new era for Windows laptops.
  • Source

Microsoft’s New AI PCs Rival Apple’s MacBooks

Microsoft revealed Copilot+ PCs, a new category of Windows PCs designed for AI. These PCs boast powerful processors, all-day battery life, and AI features like Recall for instant memory, Cocreator for image creation, Live Captions for real-time translations, and Auto Super Resolution for games.

Copilot plus PC main art
Microsoft’s New AI PCs Rival Apple’s MacBooks

The recall feature, which allows users to search and recall anything they’ve seen and interacted with on their computer screens with natural language, is especially impressive. The new PCs feature an all-new system architecture with CPU, GPU, and a high-performance Neural Processing Unit (NPU) working together. Starting at $999, Copilot+ PCs are equipped with OpenAI’s GPT-4o models.

Why does this matter?

Microsoft claims its new Arm-powered Copilot Plus PCs will outperform the MacBook Air with M3 by over 50% on sustained performance. The tech giant is advancing Arm-based computing by using AI at every level and partnering with Qualcomm. If these machines live up to the hype, they could upset the dominance of Intel-based laptops and challenge Apple’s M-series processors.

Source

🔧 TSMC and ASML machines are equipped with remote self-destruct in case of invasion

  • ASML and TSMC have the capability to disable their chipmaking machines if China invades Taiwan, according to sources.
  • ASML assured officials it can remotely shut down these machines, including its EUV units, which are crucial for producing the world’s smallest microchip transistors used in both AI and military applications.
  • US government officials, concerned about the potential impact of a Chinese invasion on the global chip supply, have discussed these security measures with both Dutch and Taiwanese counterparts.
  • Source

🧠 Microsoft’s AI chatbot will ‘recall’ everything you do on a PC

  • Microsoft introduced an upgraded version of its AI assistant, Copilot, that will remember users’ activities on their PCs to predict their next actions.
  • The new AI features, including Windows Recall, aim to enhance user experience by providing a virtual “photographic memory,” while allowing users to control their privacy settings.
  • It’s a step toward machines that “instantly see us, hear, reason about our intent and our surroundings,” said CEO Satya Nadella.
  • Source

🫠 Scarlett Johansson told OpenAI not to use her voice

  • Scarlett Johansson stated that OpenAI approached her in September about supplying her voice for GPT-4o, but she declined the offer.
  • After hearing a demo of the ChatGPT system’s “Sky” voice resembling her own, Johansson expressed shock and mentioned hiring legal counsel to address the issue with OpenAI.
  • OpenAI, while denying that Sky’s voice was meant to mimic Johansson’s, paused the use of the voice out of respect and announced plans to introduce additional voices in the future.
  • Source

Scarlett Johansson sues OpenAI for using her voice in ChatGPT 

Scarlett Johansson claims OpenAI asked her to voice ChatGPT, but she declined. Later, OpenAI released a voice named “Sky” that sounded eerily similar to her. Johansson was shocked and angered by the similarity and has hired legal counsel to investigate how the “Sky” voice was created.

OpenAI denies that the “Sky” voice was intended to resemble Johansson, has paused using it in its products, and apologizes for not communicating better. Johansson seeks transparency from OpenAI and believes that individual rights must be protected in the era of deep fakes and AI content.

Why does this matter?

Another day, another OpenAI drama. OpenAI already faces lawsuits from authors and news organizations over copyright infringement in training its AI models. Johansson’s case shows the limited legal protection actors have against AI mimicking their likeness, and some lawmakers are proposing bills to address the issue.

Source

DINO 1.5 is smarter and faster at object detection

IDEA Research launched the Grounding DINO 1.5 open-world object detection model series, with Grounding DINO 1.5 Pro for high-performance detection and Grounding DINO 1.5 Edge for efficient edge computing. Grounding DINO 1.5 Pro achieves state-of-the-art zero-shot transfer performance on several academic benchmarks, surpassing its predecessor.

DINO 1.5 is smarter and faster at object detection
DINO 1.5 is smarter and faster at object detection

The model shows strong detection capabilities across various scenarios, including common objects, long-tailed categories, dense objects, and caption phrase grounding. Grounding DINO 1.5 Pro uses a larger Vision Transformer backbone and is pretrained on the high-quality Grounding-20M dataset.

Why does this matter?

Grounding DINO 1.5 isn’t just about raw performance—it’s also incredibly versatile. Fine-tuning the model on specific datasets leads to even more impressive results, suggesting that it can lead us to make more reliable robotics, automated systems, and more while requiring less training time and data.

Source

🚚 Volvo introduces first autonomous truck

  • Volvo has unveiled its first production-ready self-driving truck, developed in partnership with the autonomous driving technology company Aurora.
  • This autonomous semi truck, based on Volvo’s VNL Class 8 model, is equipped with an array of sensors and cameras to support Aurora’s Level 4 autonomous driving system, which does not require a human driver.
  • Volvo and Aurora started collaborating in 2018, and their autonomous trucks have already logged 1.5 million miles on commercial roads; Aurora aims to deploy 20 fully autonomous trucks this year and plans to scale up to 100 trucks by 2025.
  • Source

What Else Is Happening in AI on May 21st 2024❗

💻 HP is introducing simplified laptop “AI PCs” lineup 

HP’s new “AI PCs,” the OmniBook X AI and EliteBook Ultra AI, feature Snapdragon X Elite 12-core CPUs. HP’s AI Companion software provides access to ChatGPT-3.5, while Poly Camera Pro offers AI camera controls. Moreover, the OmniBook X AI starts at $1,199.99 and the EliteBook Ultra AI at $1,699.99. (Link)

💶 Google is investing 1 billion Euros to expand its data center in Finland

This move aims to drive AI business growth in Europe. Nordic countries are popular for data centers due to cooler climate, tax breaks, and abundant renewable power. Moreover, the heat from this data center will be used to warm up local homes and buildings. Google’s Hamina data center in Finland already operates with 97% carbon-free energy and aims for net zero emissions by 2030. (Link)

🏭 Dell Technologies expands its generative AI capabilities with the Dell AI Factory

The Dell AI Factory with NVIDIA integrates hardware and software to support advanced use cases like RAG and digital assistants. Dell is also deepening partnership with Hugging Face, Meta, and Microsoft to bring open models and Azure AI services to on-premises deployments. The goal is to make it easy for enterprises to assess and implement gen AI on-premises, close to their data. (Link)

🤖 Intel’s Lunar Lake laptop processors are set to launch in Q3 2024 

These processors are designed for AI Copilot Plus PCs.  They offer 3x the AI performance of its predecessor, Meteor Lake. Intel claims Lunar Lake is 1.4 times faster in Stable Diffusion 1.5 compared to Qualcomm’s Snapdragon X Elite. The company aims to ship 40 million AI PC processors by the end of the year, with Lunar Lake featured in over 80 new laptop designs. (Link)

📜 OpenAI CEO Sam Altman apologized for the company’s strict off-boarding agreement 

This off-boarding agreement could potentially cancel departing employees’ vested equity. Meanwhile, departing execs are speaking out about the need for serious AGI safety measures. Altman and co-founder Greg Brockman defended OpenAI’s launch procedures, emphasizing their commitment to safety. (Link)

A  Daily chronicle of AI Innovations May 20th 2024:🫠 Google’s AI panic looks like Google+ fiasco 🎤 OpenAI pauses Scarlett Johansson-like voice for ChatGPT 💸 Snapchat focuses on AI with $1.5 billion yearly investment 🍏 Apple and OpenAI plan major announcement at WWDC 🤖 OpenAI’s “superalignment team,” focused on the AI risks, is no more
🚫 Sony Music warns over 700 AI companies not to steal its content 🦎 Meta’s Chameleon AI sets a new bar in mixed-modal reasoning

🫠 Google’s AI panic looks like Google+ fiasco

  • Scott Jenson, a former Google employee, criticizes the company’s AI projects as poorly motivated and driven by panic, comparing the situation to the Google+ fiasco.
  • According to Jenson, Google aims to create a Jarvis-like assistant to keep users within its ecosystem, driven by fear that competitors might get there first, similar to Apple’s strategy with Siri and OpenAI.
  • Many of Google’s AI projects revealed at I/O 2024 are still experimental or limited in availability, with uncertain value, and features like “AI Overviews” in Google Search have received criticism for potential copyright issues and misinformation.
  • Source

🎤 OpenAI pauses Scarlett Johansson-like voice for ChatGPT

  • OpenAI is pulling the ChatGPT voice known as Sky, which sounds similar to Scarlett Johansson, due to concerns about mimicking celebrities’ voices.
  • The company asserts that Sky’s voice is the natural voice of a different professional actress and not an intentional imitation of Johansson.
  • The decision to pause Sky’s use follows recent enhancements to ChatGPT’s voice mode, part of the new GPT-4o model, which aims to make the assistant more expressive and capable of reading facial expressions and translating spoken language in real-time.
  • Source

💸 Snapchat focuses on AI with $1.5 billion yearly investment

  • Snap’s CEO, Evan Spiegel, announced a major investment of $1.5 billion annually in AI and machine learning to improve Snapchat’s features and competitiveness.
  • After successfully revamping Snapchat’s advertising model, the company will now focus on collaborating with tech giants like Amazon and Google for cloud partnerships to enhance AI product innovations.
  • With growing ad revenue and new successful ad campaigns, Snapchat plans to expand content offerings and augmented reality experiences, enhancing user engagement and competing with platforms like TikTok.
  • Source

🍏 Apple and OpenAI plan major announcement at WWDC

Apple and OpenAI plan major announcement at WWDC
Apple and OpenAI plan major announcement at WWDC

  • Apple and OpenAI are planning a major joint announcement at WWDC, focusing on integrating OpenAI’s technology into iOS 18, according to Mark Gurman from Bloomberg.
  • Apple aims to improve its AI capabilities with this integration, including enhancements to Siri and features like automatic summaries of notifications and transcription of voice memos.
  • While Apple acknowledges it is behind in AI and lacks a chatbot like ChatGPT or Google Gemini, the company believes the collaboration with OpenAI is a sufficient, though temporary, solution.
  • Source

OpenAI’s “superalignment team,” focused on the AI risks, is no more

The team’s co-leads, Ilya Sutskever and Jan Leike, have resigned from OpenAI. Several other researchers from the team and those working on AI policy and governance have also left the company. Leike cited disagreements with OpenAI’s leadership about the company’s priorities and resource allocation as reasons for his departure.

image

(Source)

The team’s work will be absorbed into OpenAI’s other research efforts, with John Schulman leading research on risks associated with more powerful models.

Why does this matter?

The “superalignment” team was for ensuring the artificial general intelligence (AGI) that OpenAI claims to be working on doesn’t turn on humankind. This dismantling raises questions on the company’s commitment to AI safety and ethical standards.

Source

Sony Music warns over 700 AI companies not to steal its content

Sony Music, home to superstars like Billy Joel and Doja Cat, sent letters to over 700 AI companies and streaming platforms, warning them against using its content without permission. The label called out the “training, development, or commercialization of AI systems” that use copyrighted material, including music, art, and lyrics.

SMG recognizes AI’s potential but stresses the need to respect songwriters’ and artists’ rights. The letter asks companies to confirm they haven’t used SMG content without permission or provide details if they have.

Why does this matter?

The battle over music copyright and AI has intensified across various platforms, from YouTube’s strict rules for AI-generated music to the recent standoff between Universal Music Group and TikTok. As AI voice clones and music generation tools become more sophisticated, artists question control, compensation, and actions against copyright infringement.

Source

Meta’s Chameleon AI sets a new bar in mixed-modal reasoning

Meta AI introduces Chameleon, a family of early-fusion token-based mixed-modal models that understands and generates images and text in any order. Unlike recent foundation models that process text and images separately, Chameleon unified token space allows it to process interleaved image and text sequences.This approach allows seamless reasoning and generation across modalities.

Meta's Chameleon AI sets a new bar in mixed-modal reasoning
Meta’s Chameleon AI sets a new bar in mixed-modal reasoning

Meta researchers introduced architectural enhancements and training techniques to tackle the optimization challenges posed by this early fusion approach, including a novel image tokenizer, QK-Norm, dropout, and z-loss regularization. Remarkably, Chameleon achieves competitive or superior performance across various tasks, outperforming larger models like Flamingo-80B and IDEFICS-80B in image captioning and visual question answering despite its smaller model size.

Why does this matter?

Chameleon opens up new possibilities for more natural and intuitive human-machine interactions, similar to how we effortlessly communicate using both modalities in the real world.

Source

Trending AI Tools May 20th 2024:

  • ✍️ Paperpal – AI writing assistant offering grammar checks, paraphrasing, predictive text, and plagiarism detection. Use code RUN30 for 30% off*
  • 🗣️ ElevenLabs Audio Native – Add human-like narration to your blog or news site
  • 💻 Framer – Generate a responsive website with no code
  • 🚀 Glitter – Turn any process into a step-by-step guide
  • ⚡TestSprite Beta – Automate end-to-end software testing with AI
  • 🤖 Buffup – AI assistant that learns your intent, powered by GPT-4o

New AI Job Opportunities on May 20th 2024:

What Else Is Happening in AI on May 20th 2024❗

🤖 Google launched open-source Model Explore to visualize and debug complex AI models

It uses advanced graphics rendering techniques from the gaming industry to handle massive models. The tool offers a graphical user interface and a Python API for integration into machine learning workflows. Model Explorer lets developers identify and resolve issues quickly, especially for AI deployed on edge devices. (Link)

🇬🇧 The UK’s AI Safety Institute is opening an office in San Francisco

The institute aims to be closer to the epicenter of AI development, companies like OpenAI and Google as they are building foundational models. This new office would open this summer, giving the UK access to Silicon Valley’s tech talent and strengthening ties with the US. (Link)

📂 The EU demands Microsoft to provide internal documents on Bing’s gen AI risks

The Commission suspects Bing may have breached the Digital Services Act (DSA) due to risks like AI “hallucinations,” deep fakes, and potential voter manipulation. Microsoft has until May 27 to comply with the legally binding request for information. Failure to do so could result in fines of up to 1% of Microsoft’s total annual income or worldwide turnover. (Link)

📸 Snapchat CEO Evan Spiegel focuses on AI and ML for better UX and personalization

As its ad revenue increases, Snap plans to expand content offerings, improve recommendation algorithms, and integrate Stories with Spotlight. The company is also investing in augmented reality and sees it as a way to bring people together in shared physical environments. (Link)

😏 Researchers in the Netherlands have developed an AI sarcasm detector

The AI was trained on text, audio, and emotional content from US sitcoms, including Friends and The Big Bang Theory. The AI could detect sarcasm in unlabeled exchanges nearly 75% of the time. Further improvements could come from adding visual cues to the AI’s training data. (Link)

AI Weekly Rundown May 11-May 18th 2024: Major Breaking News from OpenAI ChatGPT-4o, Google AI Veo, Apple AI-enhanced eye tracking, Meta AI-assisted earphones , Stability AI, Humane AI

Listen here

🚫 Sam Altman quashes search engine launch rumors
📱 iOS 18 iPhones may soon have ChatGPT
💻 SoftBank’s Arm Holdings to come up with AI chips in 2025
🚀 GPT-4o: 2x faster, 50% cheaper, 5x rate limits than GPT-4 Turbo
🔍 TikTok tests ChatGPT-powered search results
🎧 Meta explores AI-assisted earphones with built-in cameras
✨ Google announced a wide array of updates across its AI ecosystem
🧠 Ilya Sutskever, OpenAI’s co-founder and chief scientist, has left the company
🍎 Apple unveils AI-enhanced eye tracking, music haptics, voice shortcuts
🤖 Humane’s AI Pin upgraded with GPT-4o
💸 Stability AI seeks buyer amid financial woes
🔄 OpenAI will bring Reddit content to ChatGPT and new products
📈 OpenAI introduces improvements to data analysis in ChatGPT
👩‍💻 GenZ and millennials optimistic about GenAI use at work: Deloitte Survey

A Daily chronicle of AI Innovations May 17th 2024: 📈 OpenAI introduces improvements to data analysis in ChatGPT 👩‍💻 GenZ and millennials optimistic about GenAI use at work: Deloitte Survey 🤝 OpenAI strikes Reddit deal to train its AI on your posts  🎵 Sony Music warns 700+ tech companies over ‘unauthorized’ use of its content to train AI 👀 The first search engine has been resurrected💸 Deepfake scam costs firm behind Sydney Opera $25 million

OpenAI introduces improvements to data analysis in ChatGPT

OpenAI introduces improvements to data analysis in ChatGPT
OpenAI introduces improvements to data analysis in ChatGPT

OpenAI is rolling out enhancements to data analysis:

  • Upload the latest file versions directly from Google Drive and Microsoft OneDrive. This allows ChatGPT to understand your Google Sheets, Docs, Slides, and Microsoft Excel, Word, and PowerPoint files more quickly
  • Interact with tables and charts in a new expandable view
  • Customize and interact with bar, line, pie, and scatter plot charts in the conversation.
  • Work on tables in real-time

These improvements will be available in OpenAI’s new flagship model, GPT-4o, for ChatGPT Plus, Team, and Enterprise users over the coming weeks.

Why does this matter?

This are exciting updates for those who use ChatGPT to work on documents for data analysis. It should save users time, give them more options and ease when using the chatbot, and make data-driven writing easier.

Source

GenZ and millennials optimistic about GenAI use at work: Deloitte Survey

Deloitte’s 2024 Gen Z and Millennial Survey connected with more than 22,800 respondents in 44 countries to explore their attitudes about work and the world around them. One of the key findings was that Gen Zs and millennials are feeling uncertain about GenAI and its potential impact on their careers.

However, respondents who frequently use GenAI at work are more likely to say they feel excitement about, and trust in the technology. Frequent users of GenAI are also more likely to believe it will free up their time, improve the way they work, and improve their work/life balance.

GenZ and millennials optimistic about GenAI use at work: Deloitte Survey
GenZ and millennials optimistic about GenAI use at work: Deloitte Survey

But, conversely, the more a respondent uses GenAI the more likely they are to have some concerns as well.

Why does this matter?

Gen AI is about to bring a huge transformation in the way we work and live. Despite concerns regarding it, GenZs and millennials are starting to think about how to adapt, whether their employers do or do not.

Source

🤝 OpenAI strikes Reddit deal to train its AI on your posts

  • OpenAI has signed a deal with Reddit to access real-time content from Reddit’s data API, similar to Reddit’s earlier $60 million deal with Google.
  • The partnership will allow Reddit to implement new AI-powered features and applications while also making OpenAI an advertising partner on the platform.
  • Details on financial terms or training data were not disclosed, and Reddit users have previously shown resistance to such changes, protesting API pricing alterations in June 2023.
  • Source

🎵 Sony Music warns 700+ tech companies over ‘unauthorized’ use of its content to train AI

  • Sony Music has sent letters to over 700 AI companies and streaming platforms, warning them against the unauthorized use of its content for AI systems, which it claims denies artists and the label control and compensation.
  • The music industry has been notably proactive in controlling the usage of its copyrighted material in AI tools, with platforms like YouTube enforcing strict rules for music content while introducing AI tools like Dream Track.
  • Universal Music Group similarly faced a dispute with TikTok over licensing, leading to a temporary removal of music by major artists like Taylor Swift and Ariana Grande until a new deal was brokered to protect artists’ rights and explore monetization opportunities involving AI.
  • Source

💸 Deepfake scam costs firm behind Sydney Opera $25 million

  • A British design firm, Arup, which is known for its work on the Sydney Opera House, lost $25.6 million in a deepfake scam involving a fake CFO and other staff members.
  • An employee in Arup’s Hong Kong office fell victim to the fraud after being convinced by realistic deepfake videos to transfer the money over 15 transactions, realizing the scam only after cross-checking with the UK office.
  • Despite the significant financial loss, Arup stated that their financial stability and business operations were unaffected, and the case is currently under police investigation.
  • Source

What Else Is Happening in AI on May 17th 2024❗

🔍Slack AI is training with customer data, but claims it is privacy-driven

To develop AI/ML models, Slack is analysing Customer Data (e.g. messages, content and files) submitted to Slack as well as Other Information (including usage information) as defined in their privacy policy and in customer agreement. But it is assuring users that their data won’t be shared across workspaces and has offered options to opt out of contributing to global models. (Link)

🤗Hugging Face is committing $10 million worth of compute to help beat big AIs

Hugging Face aims to level the playing field with AI giants like OpenAI by donating shared GPUs to the community through a new program called ZeroGPU. The shared GPUs are accessible to multiple users or applications concurrently, eliminating the need for each user or application to have a dedicated GPU. (Link)

☁️Microsfot to offer its cloud customers AMD alternative to Nvidia AI processors

Microsoft will give its cloud customers a platform of AMD AI chips that will compete with components made by Nvidia, with details to be given at its Build developer conference next week. It will also launch a preview of its new Cobalt 100 custom processors at the conference. (Link)

🤝Ampere teams up with Qualcomm to launch an Arm-based AI server

The two companies are combining forces to to offer an AI-focused server that uses Ampere’s CPUs and Qualcomm’s Cloud AI 100 Ultra AI inferencing chips for running– not training– models. (Link)

📰Gannett is adding AI-generated summaries to the top of its articles

Gannett, the media company that owns hundreds of newspapers in the US, is launching a new program that adds AI-generated bullet points at the top of news stories. The summaries appear to be already live and are aimed to enhance the reporting process and elevate the audience experience. (Link)

A Daily chronicle of AI Innovations May 16th 2024: 🤔 Microsoft’s AI ambitions lead to 30% rise in emissions🪫 Stability AI reportedly runs out of money and in negotiations to sell 🍎Apple unveils AI-enhanced eye tracking, music haptics, voice shortcuts 🤖Humane’s AI Pin upgraded with GPT-4o 💸Stability AI seeks buyer amid financial woes 📽️Mozilla Firefox incorporates NVIDIA RTX video to improve video quality on Windows ☁️Musk’s AI firm xAI In talks with Oracle for $10 billion cloud deal ❄️Snowflake invests in Metaplane to enhance data quality 😱 Ilya Sutskever and Jan Leike leave OpenAI 📱 Android phones enter the AI era

 Microsoft’s AI ambitions lead to 30% rise in emissions

  • Microsoft’s commitment to being carbon-negative by 2030 is threatened by a 30% increase in its carbon emissions, largely due to its growing focus on AI technologies.
  • The company’s expansion of AI operations necessitates building more data centers, which rely on energy-intensive materials like steel, concrete, and microchips, further complicating their carbon reduction goals.
  • Despite its renewable energy claims, Microsoft is acknowledged for using instruments like renewable energy credits (RECs) that might mask actual carbon emissions, indicating a challenging path ahead to meet its sustainability targets.
  • Source

🪫 Stability AI reportedly runs out of money and in negotiations to sell

  • Stability AI, known for its open-source image generator Stable Diffusion, is negotiating a sale due to financial struggles, despite a valuation of $1 billion.
  • The company raised over $100 million in 2022 but struggled to turn its free software into revenue while facing stiff competition from rivals like OpenAI and Midjourney.
  • Founder and CEO Emad Mostaque resigned amid management issues, and the company faces significant expenses and liabilities, projecting $153 million in costs against only $11 million in revenue for 2023.
  • Source

Apple unveils AI-enhanced eye tracking, music haptics, voice shortcuts

Apple recently unveiled new accessibility features that will launch later this year. One of these features is Eye Tracking, which enables users with physical disabilities to navigate their iPhones and iPads using only their eyes.

By leveraging the front-facing camera and AI, Eye Tracking allows precise control through a “Dwell Control” feature, eliminating the need for additional hardware.

Another noteworthy addition is Music Haptics, designed for users who are deaf or hard of hearing. When enabled, the Taptic Engine in iPhones plays taps, textures, and refined vibrations synchronized with music audio.

Additionally, Apple has introduced Voice Shortcut, which allows users to create custom voice commands for specific actions on their devices. With Voice Shortcut, users can perform tasks hands-free by simply speaking their personalized commands

Why does this matter?

This would encourage the development and advancement of AI technologies in the accessibility domain, ultimately benefiting individuals with disabilities across different platforms and devices.

Source

Humane’s AI Pin upgraded with GPT-4o

Humane has upgraded its popular AI Pin product with the GPT-4o model. This upgrade aims to enhance the AI Pin’s performance. The new version runs faster, provides more accurate answers, and offers richer information.

Humane’s AI Pin upgraded with GPT-4o
Humane’s AI Pin upgraded with GPT-4o

Why does this matter?

As more sophisticated AI models become accessible, we can expect further improvements in various applications, from digital assistants to natural language understanding and generation.

Source

😱 Ilya Sutskever and Jan Leike leave OpenAI

Ilya Sutskever and Jan Leike leave OpenAI
Ilya Sutskever and Jan Leike leave OpenAI

OpenAI co-founder and chief scientist Ilya Sutskever announced that he is leaving the company — following months of speculation of Sutskever’s role from the November 2023 Sam Altman ousting.

  • Sutskever said he is confident that OpenAI will ‘build AGI that is both safe and beneficial’ under the current leadership.
  • Also leaving is Sutskever’s superalignment group co-lead Jan Leike, announcing his departure with a cryptic resignation post.
  • The news follows months of departures largely coming from OpenAI’s superalignment and safety teams, fueling speculation.
  • OpenAI CEO Sam Altman named Jakub Pachocki as the new chief scientist, a key researcher on the creation of GPT-4.

After months of tensions, the speculation around Ilya’s future with Sam is finally put to rest. But questions surrounding the safety team departures still remain. Also important to watch is where Sutskever and Leike land next, with two of AI’s brightest minds now officially on the market.

📱 Android phones enter the AI era

Android phones enter the AI era
Android phones enter the AI era

Google announced a host of new AI integrations coming to Android phones at its I/O Developer Conference, bringing its powerful Gemini model on-device to enable upgraded smartphone experiences.

  • Google’s Gemini Nano model will be integrated into the Pixel later this year, allowing for enhanced multimodal capabilities.
  • Gemini features will be easily accessible with a new overlay, which improves and understand context to provide dynamic suggestions.
  • A Circle to Search feature, allowing users to query anything on screen, gains homework help features via a LearnLM model.
  • Google is also bringing a new AI security feature soon, providing real-time alerts on calls that appear to be scams.

While everyone awaits the iPhone AI announcements, Google’s Android AI era is rolling. With Gemini coming directly on-board, the potential is there — but if Apple integrates ChatGPT, it may remain tough sledding against the dominant market leader.

Source

AI TRAINING 🗣️ How to clone your voice using AI

How to clone your voice using AI
How to clone your voice using AI

A new model on Replicate called OpenVoice lets you clone any voice for free with just an audio file and the desired text to be turned into speech.

Step-by-step:
  1. Access OpenVoice on Replicate here and log in with your GitHub account.
  2. Upload the audio file of the voice you want to clone where it says ‘audio’. The longer, the better.
  3. Fill the ‘Text’ field with the text you want to convert into a speech.
  4. Click on the ‘Run’ button and listen/download the generated audio with the cloned voice!

New AI Job Opportunities on May 16th 2024

  • 🔬 Lambda – Manufacturing Quality Engineer
  • 📋 Anthropic – People Operations Coordinator
  • 🤖 Figure AI – Humanoid Robot Pilot
  • 💻 Luma AI – Senior Software engineer- Data infrastructure

What Else Is Happening in AI on May 16th 2024❗

🧑‍Instagram co-founder joins Anthropic as head of product

Anthropic has appointed Mike Krieger, the co-founder of Instagram and Artifact, as head of product. In this pivotal role, Krieger will steer the company’s product engineering, management, and design efforts as Anthropic expands its suite of AI apps and aims to bring Claude to a wider audience. (Link)

💼 IBM plans 800 new jobs in Ireland

IBM plans to hire up to 800 more people in Ireland to help build advanced software underpinned by AI technology. This is the biggest job announcement in the country’s multinational sector this year. The roles will be in research and development, digital sales, and consulting.  (Link)

📽️Mozilla Firefox incorporates NVIDIA RTX video to improve video quality on Windows

Mozilla Firefox has incorporated NVIDIA RTX Video, a technology that leverages AI to improve video quality on Windows PCs and workstations. This brings a host of AI-powered enhancements to online videos, including upscaling and HDR, for a more immersive viewing experience. Users can easily access these features by updating GeForce RTX drivers and activating Windows HDR.  (Link)

☁️Musk’s AI firm xAI In talks with Oracle for $10 billion cloud deal

Elon Musk’s AI startup, xAI, is in talks with Oracle executives about a $10 billion deal to rent cloud servers from Oracle over several years. This move comes as Musk’s company seeks to keep pace with rivals like OpenAI, Microsoft, and Google to dominate the AI field. (Link)

❄️Snowflake invests in Metaplane to enhance data quality 

Snowflake, a data cloud platform, has invested in Metaplane, a Boston-based startup. Metaplane focuses on helping enterprises identify and rectify data quality issues using an end-to-end AI-powered platform. The investment aims to enhance integration between Metaplane’s data observability offering and the Snowflake data cloud. (Link)

A Daily chronicle of AI Innovations May 15th 2024: 🤖Google unveiled the ‘future of AI’ at I/O event 🧪OpenAI chief scientist Ilya Sutskever is leaving   🚗Former Cruise CEO starts robotics firm with ex-Tesla AI manager ✨ Google announced a wide array of updates across its AI ecosystem 🧠Ilya Sutskever, OpenAI’s co-founder and chief scientist, has left the company 🏖️ Expedia is launching AI features, including a bot named Romi 🧬 NVIDIA and Recursion have collaborated to build BioHive-2, an AI supercomputer for drug discovery  🚀 NASA appoints David Salvagnini, its first chief AI officer  💰 SoftBank plans to invest $75-$150 million in Indian data center and industrial robotics 💶 Microsoft announces €4 billion investment in France to accelerate AI adoption 🎥 Google’s Gemini updates and Sora competitor🔎 Google’s AI agents and AI search upgrades

🎥 Google’s Gemini updates and Sora competitor

Google’s Gemini updates and Sora competitor
Google’s Gemini updates and Sora competitor

Google just kicked off its I/O Developer’s Conference, announcing a wide array of updates across its AI ecosystem — including enhancements across its flagship Gemini model family and a new video generation model to rival OpenAI’s Sora.

Gemini model updates:

  • New updates to 1.5 Pro include a massive 2M context window extension and enhanced performance in code, logic, and image understanding.
  • Gemini 1.5 Pro can also utilize the long context to analyze a range media types, including documents, videos, audio, and codebases.
  • Google announced Gemini 1.5 Flash, a new model optimized for speed and efficiency with a context window of 1M tokens.
  • Gemma 2, the next generation of Google’s open-source models, is launching in the coming weeks, along with a new vision-language model called PaliGemma.
  • Gemini Advanced subscribers can soon create customized personas called ‘Gems’ from a simple text description, similar to ChatGPT GPTs.

Video and image model upgrades:

  • Google revealed a new video model called Veo, capable of generating over 60-second, 1080p resolution videos from text, image, and video prompts.
  • The new Imagen 3 text-to-image model was also unveiled with better detail, text generation, and natural language understanding than its predecessor.
  • VideoFX text-to-video tool, featuring storyboard scene-by-scene creation and the ability to add music to generations.
  • VideoFX is launching in a ‘private preview’ in the U.S. for select creators, while ImageFX (with Imagen 3) is available to try via a waitlist.

Why it matters: Gemini’s already industry-leading context window gets a 2x boost, enabling endless new opportunities to utilize AI with massive amounts of information. Additionally, Sora officially has competition with the impressive Veo demo — but which one will make it to public access first?

🔎 Google’s AI agents and AI search upgrades

Google’s AI agents and AI search upgrades
Google’s AI agents and AI search upgrades

Google just showcased its new AI agent project ‘Project Astra’, alongside a slew of updates to infuse AI across search and enable Gemini to reason and take more advanced actions for users.

Progress on AI agents:

  • Google announced Project Astra, a real-time AI agent prototype that can see, hear, and take actions on a user’s behalf.
  • The demo showcased a voice assistant responding to what it sees and hears, including code, images, and video — capable of advanced reasoning and recall.
  • Public access for Astra is expected through the Gemini app later this year.
  • Google also showed off ‘AI teammates’, agents that can answer questions on emails, meetings, and other data within Workspace.
  • Live is also rolling out in the coming months, allowing users to speak and converse with Gemini in near real-time.

Search upgrades:

  • Google Search now features expanded AI Overviews, advanced planning capabilities, and AI-organized search results.
  • Gemini will be able to execute more complex planning, such as planning, maintaining, and updating trip itineraries.
  • Search will also receive ‘multi-step reasoning’ capabilities, allowing Gemini to break down questions and speed up research.
  • Users can also now ask questions with video, allowing Search to analyze visual content and provide helpful AI Overviews.

Why it matters: We officially have a new voice assistant battle — with OpenAI and Google both showcasing mind-blowing new capabilities in just the last two days alone. Also, despite rumblings of an OpenAI search product and excitement over platforms like Perplexity, it’s going to be difficult to unseat the king of search. Especially as they integrate advanced AI across the entire ecosystem in an impressive fashion.

🚗 Former Cruise CEO starts robotics firm with ex-Tesla AI manager

  • Former Cruise CEO Kyle Vogt, who resigned in November following a self-driving car accident, has launched a new robotics firm called The Bot Company with former Tesla AI Manager Paril Jain.
  • The Bot Company aims to develop robots that handle everyday chores to give people more free time, and has already raised $150 million from investors.
  • The announcement of The Bot Company comes amid significant challenges for Cruise, which had its self-driving operations suspended by the California DMV following a pedestrian accident.
  • Source

Google announced a wide array of updates across its AI ecosystem

Major headlines were Project Astra, an AI agent that can see and hear users in real time, and AI teammates that can answer questions and organize data within Workspace.

Google also introduced Veo, a high-quality video generation model, and significant updates to Google Search, including expanded AI Overviews, advanced planning capabilities, and AI-organized search results.

The company launched Gemini 1.5 Pro, boasting a massive 2M context window extension, and Imagen 3, the next version of its text-to-image model. Other notable announcements included:

  • Gemma 2 updates: New 27-billion-parameter model launching in June
  • Google Play: New app discovery feature and developer tools
  • Android feature to detect potential scams during calls using Gemini Nano
  • Ask Photos: AI-powered search in Google Photos using natural language queries
  • Gemini in Gmail for searching, summarizing, and drafting emails
  • Gemini 1.5 Pro: Increased input capacity to 2 million tokens
  • Gemini Live: In-depth voice chats with Gemini on smartphones
  • Gemini Nano: Smallest AI model built into Chrome desktop client
  • Gemini on Android: Deep integration with Android OS and Google apps
  • Gemini on Google Maps: Generative AI summaries for places and areas
  • Tensor Processing Units (TPU): Sixth-generation Trillium chips with 4.7x performance boost
  • Project IDX: AI-centric browser-based development environment in open beta
  • Circle to Search: AI-powered feature for instant answers using gestures on Android
  • Pixel 8a: New smartphone with Tensor G3 chip, starting at $499
  • Pixel Slate: Google’s Pixel Tablet is now available with or without the base

With 22 announcements, Google is making everyone a bit overwhelmed. By integrating AI across its vast ecosystem, Google aims to provide users with more personalized and innovative experiences. But Google Project Astra and Veo are not available to the public yet. This may irritate customers compared to OpenAI, which launches new products that customers can play with immediately.

Source

Ilya Sutskever, OpenAI’s co-founder and chief scientist, has left the company

Ilya Sutskever, OpenAI’s co-founder and longtime chief scientist, has left the company. OpenAI CEO Sam Altman announced the news on X, expressing his sadness and gratitude for Sutskever’s contributions. Jakub Pachocki, OpenAI’s research director, will step up as a chief scientist.

Sutskever’s departure comes amidst reports of disagreements with Altman over OpenAI’s direction, mainly concerns about rushing AI-powered product launches at the expense of safety. The situation escalated last November when Sutskever and OpenAI CTO Mira Murati approached the company’s previous board to express their concerns, leading to a brief attempt to fire Altman. However, Altman was swiftly reinstated, and much of the old board resigned.

Why does it matter?

As one of the most accomplished minds in AI, Sutskever’s departure raises questions about the future of OpenAI’s approach to AI development and safety. The incident also highlights the growing tensions within the AI industry between rapidly commercializing AI technologies and the need for responsible development and deployment practices.

Source

What Else Is Happening in AI on May 15th 2024❗

🏖️ Expedia is launching AI features, including a bot named Romi

It helps users search for hotels, build itineraries, and make changes via iMessage and WhatsApp. The company also introduces smart search, allowing travelers to find hotels based on desired features. Expedia pulls data from AccuWeather and Yelp to tailor search results and provide real-time updates. (Link)

🧬 NVIDIA and Recursion have collaborated to build BioHive-2, an AI supercomputer for drug discovery

BioHive-2 features 504 NVIDIA H100 GPUs, delivering 2 exaflops of AI performance, nearly 5 times faster than its predecessor. Powered by Recursion’s massive 50-petabyte dataset and AI models like Phenom, these companies aim to simulate biology and fast-track the identification of promising drug candidates. (Link)

🚀 NASA appoints David Salvagnini, its first chief AI officer

The move aims to ensure NASA stays ahead in AI advancement and responsible use. Salvagnini will lead NASA’s efforts in guiding the responsible use of AI and collaborating with other agencies, academia, and industry partners. (Link)

💰 SoftBank plans to invest $75-$150 million in Indian data center and industrial robotics

The move aligns with the Japanese tech giant’s global strategy to capitalize on the power of artificial intelligence as it shifts focus from e-commerce and fintech to high-growth industries like logistics and robotics. The potential investments mark the end of a two-year hiatus in SoftBank’s deal-making activities in India. (Link)

💶 Microsoft announces €4 billion investment in France to accelerate AI adoption

The tech giant will expand its cloud and AI infrastructure, launch skilling programs, and accelerate French startups through its new Microsoft GenAI Studio, positioning France as a leader in the AI revolution. It aims to train 1 million people and support 2,500 startups by 2027. (Link)

AI TRAINING: 🤖 Use ChatGPT’s highlighting for context

AI TRAINING: 🤖 Use ChatGPT's highlighting for context
AI TRAINING: 🤖 Use ChatGPT’s highlighting for context

ChatGPT now allows you to highlight parts of its responses for quick follow-up questions, partial response rewrites, reusing old context, and more.

Step-by-step:

  1. Prompt ChatGPT and generate a response.
  2. Highlight relevant parts of the response you want to follow-up on and click the double quote icon above the highlighted text.
  3. The highlighted text will be automatically added in the next prompt so you can ask for clarification, rewrites, counterpoints, and more.

New AI Job Opportunities on May 15th 2024

  • 📣 OpenAI – Media Relations, Policy Communications
  • 📋 Lambda – Sourcing Manager
  • 🏢 C3 AI – General Manager, Federal Systems
  • 💼 Glean – Accounting Manager

🤖 Google unveiled the ‘future of AI’ at I/O event

  • Google I/O just ended and a lot of announcements were made. Gemini 1.5 Pro will increase its context window from one to two million tokens and a new model called Gemini Flash was announced, which is optimized for speed and efficiency
  • The company launched Astra, a multimodal AI assistant for everyday life. It can process text, video, and audio in real time. In a video, Google showed Astra identifying speakers, crayons and other objects in response to a camera image and voice commands.
  • Google also unveiled its latest AI models for creating media content: Veo, for creating 1080p videos, and Imagen 3, for generating images from text descriptions.
  • Source

🧪 OpenAI chief scientist Ilya Sutskever is leaving

  • Ilya Sutskever, OpenAI’s co-founder and chief scientist, is officially leaving the company after his involvement in the failed attempt to remove CEO Sam Altman and subsequently changing his stance.
  • Sam Altman announced that Jakub Pachocki, who has led significant projects such as GPT-4 and OpenAI Five, will take over as the new Chief Scientist at OpenAI, ensuring the company’s continued progress towards its mission.
  • Jan Leike, who has been leading the Superalignment team aimed at controlling more powerful AI, has also resigned, with his responsibilities now being taken over by OpenAI co-founder John Schulman.
  • Source

A Daily chronicle of AI Innovations May 14th 2024: 🚀 GPT-4o: 2x faster, 50% cheaper, 5x rate limits than GPT-4 Turbo 🗣️ OpenAI unveils GPT-4o and new voice capabilities 🔍 TikTok tests ChatGPT-powered search results 🎧 Meta explores AI-assisted earphones with built-in cameras 💰 Anthropic launches Claude in Europe in the hope of raising more money 🤖 UAE launches a new AI model to compete with big tech 🖥️ Japan team builds Gen AI foundation model using supercomputer Fukagu  ⏰ Microsoft Places finds the best time for your next office day 🤝 U.S. and China to meet in Geneva to discuss AI risks 🎧 Meta developing AI-powered ‘Camerabuds’ 🥴 Meta’s AI system ‘Cicero’ learning how to lie, deceive humans

🗣️ OpenAI unveils GPT-4o and new voice capabilities

OpenAI just unveiled GPT-4o, a new advanced multimodal model that integrates text, vision and audio processing, setting new benchmarks for performance – alongside a slew of new features.

The new model:

  • GPT-4o provides improved performance across text, vision, audio, coding, and non-English generations, smashing GPT-4T’s performance.
  • The new model is 50% cheaper to use, has 5x higher rate limits than GPT-4T, and boasts 2x the generation speed of previous models.
  • The new model was also revealed to be the mysterious ‘im-also-a-good-gpt2-chatbot’ found in the Lmsys Arena last week.

Voice and other upgrades:

  • New voice capabilities include real-time responses, detecting and responding with emotion, and combining voice with text and vision.
  • The demo showcased feats like real-time translation, two AI models analyzing a live video, and using voice and vision for tutoring and coding assistance.
  • OpenAI’s blog also detailed advances like 3D generation, font creation, huge improvements to text generation within images, sound effect synthesis, and more.
  • OpenAI also announced a new ChatGPT desktop app for macOS with a refreshed UI, integrating directly into computer workflows.

Free for everyone:

  • GPT-4o, GPTs, and features like memory and data analysis are now available to all users, bringing advanced capabilities to the free tier for the first time.
  • The GPT-4o model is currently rolling out to all users in ChatGPT and via the API, with the new voice capabilities expected to arrive over the coming weeks.

Source

GPT-4o: 2x faster, 50% cheaper, 5x rate limits than GPT-4 Turbo

OpenAI unveiled its new flagship model, GPT-4o (“o” stands for “omni”). It provides GPT-4-level intelligence but is 2x faster, 50% cheaper, has 5x higher rate limits, and enhanced text, voice, and vision capabilities than GPT-4 Turbo. It also matches GPT-4 Turbo performance on text in English and code, with significant improvements for text in non-English languages.

GPT-4o can respond to audio inputs in as little as 232 ms, with an average of 320 ms, which is similar to human response time in a conversation. It is available in the ChatGPT free tier and to Plus users.

Why does it matter?

GPT-4o is a step towards much more natural human-computer interaction. It accepts any combination of text, audio, and image as input and generates any combination of text, audio, and image outputs. GPT-4o will set a new benchmark for AI capabilities and pave the way for more intelligent and accessible AI systems.

Source

TikTok tests ChatGPT-powered search results

TikTok is testing a new search feature that uses AI to generate results for some queries. For certain search terms, like recipes or product recommendations, TikTok will display AI-generated answers at the top of the results. These answers are powered by ChatGPT, and when you click on one of these AI results, it takes you to a page with a full response.

TikTok tests ChatGPT-powered search results
TikTok tests ChatGPT-powered search results

TikTok is also testing “search highlights,” which appear at the top of some searches. However, it’s unclear whether these highlights are AI-generated.

Why does it matter?

It could make search on TikTok more authoritative and reliable by providing factual answers from AI models like ChatGPT. This could greatly impact how people use TikTok for search and information gathering, thereby positioning TikTok more directly as a search engine competitor to Google, catering to younger users’ preferences.

Source

Meta explores AI-assisted earphones with built-in cameras

Meta is currently exploring AI-powered earphones with built-in cameras. The project is known internally as “Camerabuds;” however, it’s uncertain if the final product will be in-ear earbuds or over-ear headphones. The cameras let the AI take in surrounding information visually, allowing it to translate foreign languages, identify objects, and more.

CEO Mark Zuckerberg has reportedly seen several possible designs but has not been satisfied with them. Outside of design, the major hurdle is fitting a camera and batteries in a tiny device like an earbud.

Why does it matter?

For the AI wearable space, Meta’s development hints at a growing interest in AI-infused devices, potentially paving the way for more sophisticated and interactive wearables. Regarding earphones, this development suggests a potential shift towards more intelligent and feature-rich earphone designs, offering users enhanced functionalities like real-time translation and object identification.

Source

🎧 Meta developing AI-powered ‘Camerabuds’

Meta is reportedly in the early stages of developing AI-powered earphones, known internally as “Camerabuds,” —  aiming to compete with OpenAI and Apple as tech giants rush to infuse AI into wearable devices.

  • Camerabuds’ would map user surroundings, capable of identifying objects and translating foreign languages using built-in cameras.
  • Meta already has its AI-powered Ray Ban smart glasses, while OpenAI and Apple are also exploring similar AI wearable earbud tech.
  • Potential challenges include bulkiness, heat generation, and privacy concerns, especially for users with long hair that might obstruct the cameras.

Despite Meta’s shaky track record with hardware ventures, Mark Zuckerberg is investing heavily in a future that he believes includes AI embedded into every device. But will standalone devices like this be able to win over users if and when a fully AI-integrated phone hits the market?

Source

🥴 Meta’s AI system ‘Cicero’ learning how to lie, deceive humans

  • Meta’s AI, Cicero, has learned to lie and deceive humans, excelling in the strategy game Diplomacy by betraying alliances.
  • The study, published in the journal Patterns, indicated that several AI systems, including Cicero, AlphaStar, and Pluribus, leverage deception to improve performance in competitive tasks and games.
  • Experts warned that advanced AI’s deceptive abilities pose increasing risks to society, suggesting that deceptive AI systems should be classified as high risk if outright banning them is not feasible.
  • Source

New AI Job Opportunities on May 14th, 2024

What Else Is Happening in AI on May 14th, 2024❗

💰 Anthropic launches Claude in Europe in the hope of raising more money

Anthropic launches Claude in Europe with support for “multiple languages,” including French, German, Italian, and Spanish across Claude.ai, its iOS app, and its business plan for teams. The launch comes after Anthropic extended its API to Europe for developers. Both moves are part of its bigger push to raise more money for rapid growth. (Link)

🤖 UAE launches a new AI model to compete with big tech

Abu Dhabi’s Technology Innovation Institute (TII) released Falcon 2 series: Falcon 2 11B, a text-based model, and Falcon 2 11B VLM, a vision-to-language model that can generate a text description of an uploaded image. This Gen AI model could compete with some of the biggest technology companies and make UAE a strong candidate in the Gen AI space. (Link)

🖥️ Japan team builds Gen AI foundation model using supercomputer Fukagu

Tokyo Institute of Technology researchers and others have announced the development of an LLM that can serve as a foundation for Gen AI using the Japanese supercomputer Fugaku. Trained extensively on Japanese data, which accounts for 60% of the total training data, the Fugaku-LLM model is expected to lead research on Gen AI tailored to domestic needs. (Link)

⏰ Microsoft Places finds the best time for your next office day

Microsoft has launched a new app called Microsoft Places that uses AI to help employees and managers coordinate their in-office days and book meeting rooms. The application integrates with Outlook and Microsoft Teams, allowing employees to view and share their office plans and get AI-powered suggestions on which days they should come in. (Link)

🤝 U.S. and China to meet in Geneva to discuss AI risks

The U.S. and China will hold their first formal government-to-government discussion on the risks of artificial intelligence in Geneva. The U.S. side will be led by officials from the National Security Council and State Department, while the Chinese side will be led by officials from the Foreign Ministry and the National Development and Reform Commission. (Link)

A Daily chronicle of AI Innovations May 13th 2024: 🔮 Rumor mill: OpenAI’s Monday reveal; Hello GPT-4o; OpenAI Livestream event; Sam Altman – “No Fixed Timeline for GPT5”; Apple Close to Agreement with OpenAI to Bring ChatGPT to iPhone; 🛠️ Anthropic’s new tool automates 🚫 Sam Altman quashes search engine launch rumors📱 iOS 18 iPhones may soon have ChatGPT 💻 SoftBank’s Arm Holdings all set to come up with AI chips in 2025 prompting;  🎥 How to create realistic AI avatar videos 🧒Anthropic enables safe AI technology for kids 🚗SoundHound AI and Perplexity partner to enhance voice assistants for cars and IoT devices ⚒️U.K. agency releases tools for testing AI model safety 💻IBM Open-Sources Granite Code Models for Easier Coding 🎓Hong Kong University brings Albert Einstein to life using AI as a virtual lecturer 

🔮 Rumor mill: OpenAI’s Monday reveal

OpenAI is set to demo new features and updates to ChatGPT and GPT-4 today at 10 AM PT, with new speculation including a ‘Her’ style voice assistant with both audio and visual capabilities.

  • According to The Information, OpenAI’s demo will include a virtual assistant with visual AND audio understanding.
  • The report also claims the new reveal might have the ability to make ‘existing voice assistants like Siri more useful.’
  • CEO Sam Altman shot down rumors of a new search engine competitor and GPT-5, but said the reveal is something that ‘feels like magic’.
  • Additional speculation includes the ability to initiate and receive phone calls inside of ChatGPT.
  • Apple and OpenAI are also reportedly ‘closing in’ on a deal to incorporate ChatGPT into iOS 18.

🤖 OpenAI releases GPT-4o, a faster model that’s free for all

  • OpenAI announced the launch of GPT-4o, an improved version of its GPT-4 model, promising faster performance and enhanced capabilities in text, vision, and audio.
  • GPT-4o is freely available to all ChatGPT users, with paid users benefiting from up to five times the capacity limits of free users. In OpenAI’s API, GPT-4o is 50% faster than GPT-4 (specifically GPT-4 Turbo), 50% cheaper and has higher rate limits.
  • The model is described as “natively multimodal,” capable of generating content or interpreting commands across voice, text, or images.
  • Source

🕒 Microsoft Places uses AI to find the best time for your next office day

  • Microsoft has launched a new AI-powered app called Microsoft Places, designed to facilitate better in-office coordination for hybrid and flexible work setups by integrating with Microsoft 365.
  • The app allows users to plan and share their office days, integrates location data into Outlook calendars for scheduling, and is set to include features in Microsoft Copilot for automated scheduling recommendations and adjustments.
  • Microsoft Places will also offer a space analytics feature for real estate management, improve room booking capabilities in Outlook, and feed into Teams to enhance real-time location awareness among colleagues.
  • Source

🔍 Wikipedia’s new Chrome extension fact-checks the web with ChatGPT

  • The Wikimedia Foundation has developed an experimental Chrome extension named “Citation Needed” that utilizes ChatGPT and Wikipedia to check the veracity of online content.
  • By integrating ChatGPT, the extension cross-references user-selected text with Wikipedia entries and citations, offering feedback on the accuracy and reliability of the information.
  • This initiative aims to gauge public interest in factual verification using Wikipedia, explore Wikipedia’s credibility as a source, and assess the potential of AI in enhancing content verification processes.
  • Source

🛠️ Anthropic’s new tool automates prompting

Anthropic just launched a new Prompt Generator tool for its business and API users, helping to automatically craft optimal prompts via natural language when completing tasks with its Claude models.

  • The generator leverages advanced prompt techniques like chain-of-thought reasoning for more ‘effective, precise, and reliable’ outputs.
  • Console users can also test prompt performance via dynamic variable insertion, optimizing prompts based on various situations.
  • Anthropic released a Prompt Library earlier this year, featuring a range of optimized prompts that users can copy and paste.

 While ‘Prompt Engineer’ was a popular term thrown around as a potential future job, the reality is that AI can help simplify the task with optimal prompts that it creates on its own. While Anthropic’s tool is only on the API side for now, it’s only a matter of time before similar features make their way to the full consumer side.

Sam Altman quashes search engine launch rumors

OpenAI’s CEO, Sam Altman, denied rumors of launching a search engine ahead of Google I/O, stating that the upcoming announcement is neither related to GPT-5 nor a search engine. He described the forthcoming revelation as something that “feels like magic.”

Sam Altman quashes search engine launch rumors
Sam Altman quashes search engine launch rumors

However, there are reports that OpenAI is developing an AI voice assistant.

Why does this matter?

OpenAI has a track record of pushing the boundaries of AI capabilities, so any new development from them is likely to attract attention and disrupt the status quo.

As for reports of OpenAI entering into the voice assistant domain, if true, we might see OpenAI disrupt one more niche. With companies like Amazon (Alexa), Google (Google Assistant), and Apple (Siri) dominating the market, the competition might be stiff this time around.

Source

iOS 18 iPhones may soon have ChatGPT

OpenAI will bring ChatGPT to iPhones. The integration is expected to happen with the upcoming iOS 18. This could significantly enhance the user experience and expand the capabilities of Apple’s virtual assistant, Siri.

Parallel to negotiations with OpenAI, the tech giant is also engaging with Google’s parent company, Alphabet, to integrate the Gemini chatbot into Apple’s devices. These ongoing talks, however, haven’t led to an agreement yet.

Why does this matter?

Integrating ChatGPT into iOS 18 iPhones could bolster Siri’s capabilities, offering users a more versatile AI assistant experience. With ChatGPT handling various queries directly within iOS, users won’t need to rely on third-party apps for assistance, streamlining their interactions and enhancing their overall experience.

Source

SoftBank’s Arm Holdings all set to come up with AI chips in 2025

SoftBank’s subsidiary, Arm, plans to develop AI chips and launch them in 2025. Arm is establishing a dedicated AI chip division and targeting a prototype by spring 2025 and mass production by autumn. SoftBank will contribute to the initial cost of 10 trillion yen ($64 billion).

The AI chip business could be spun off to SoftBank. Negotiations with Taiwan Semiconductor Manufacturing Corp (TSMC) and others for production capacity are underway. The move reflects rising demand for AI chips, which has boosted Arm’s share price since its IPO in September.

Why does this matter?

As AI applications expand across industries, there’s a demand for efficient and optimized hardware solutions. Arm’s entry into the AI chip market could disrupt the current landscape dominated by companies like NVIDIA.

Source

AI Tutorial: 🎥 How to create realistic AI avatar videos

  1. Head over to HeyGen’s website and sign up for free.
  2. Click where it says ‘AI Studio’ on the left bar of the Dashboard.
  3. Choose a given template by selecting ‘Templates’ or create one from scratch by pressing ‘Create with AI Studio’.
  4. In the Studio, you can add and modify any part of your presentation. For example, you can add text and images, change the background, select an AI avatar and its voice, create a script using AI, and more.
  5. When your video is ready, press “Submit” and check out the final video!

AI Research: 🎓 MIT studies AI deception capabilities

MIT studies AI deception capabilities
MIT studies AI deception capabilities

MIT researchers just published a new study detailing AI’s ability to engage in deceptive behaviors like lying, bluffing, and disguising true capabilities in testing.

  • The analysis found numerous examples of AI deception, such as Meta’s Cicero Diplomacy bot telling premeditated lies to players.
  • Other cases included negotiation systems misrepresenting preferences, AI bluffing in poker, and ‘playing dead’ to fool safety tests.
  • Growing deceptive capabilities could influence societal dangers like election tampering, fraud, and general control over AI systems.

The combination of AI’s superhuman persuasiveness and the growing ability to engage in deception are a reminder of a potential dark side to the tech’s advancement. As these qualities emerge, the importance of reliably aligning and better understanding model behavior only grows.

New AI Job Opportunities on May 13th, 2024

What Else Is Happening on May 13th, 2024❗

🧒Anthropic enables safe AI technology for kids

Anthropic now allows minors to use its generative AI systems through third-party apps. Developers must implement safety measures, including age verification and content moderation. It also plans to periodically audit these apps for compliance with child safety and data privacy regulations, including the Children’s Online Privacy Protection Act (COPPA).  (Link)

🚗SoundHound AI and Perplexity partner to enhance voice assistants for cars and IoT devices

Perplexity’s online LLM capabilities have been integrated into SoundHound’s Chat AI, enabling conversational responses with real-time knowledge sourced from the web. This collaboration aims to provide more accurate and comprehensive answers to user queries.  (Link)

⚒️U.K. agency releases tools for testing AI model safety

The toolset, called Inspect, is an open-source platform that assists industry, research organizations, and academia in evaluating the capabilities of AI models, such as core knowledge and reasoning abilities. It aims to strengthen AI safety and is the first platform of its kind backed by a state-supported body. (Link)

💻IBM Open-Sources Granite Code Models for Easier Coding

IBM Research has recently released a family of Granite code models to the open-source community. These models aim to make coding easier for developers by assisting in writing, testing, debugging, and shipping reliable software. They find use in diverse scenarios, from application modernization to memory-constrained environments like IBM’s WCA for Z. Performance evaluations show superiority over alternatives like Mistral-7B and LLama-3-8B. (Link)

🎓Hong Kong University brings Albert Einstein to life using AI as a virtual lecturer

The Hong Kong University of Science and Technology (HKUST) is experimenting with AI-generated lecturers, featuring personalities like Albert Einstein and even popular anime characters. The goal is to address the global teacher shortage while embracing AI’s transformative power in education. (Link)

🧠 Google’s new Med-Gemini-3D model can read 3D scans like head CTs, answer questions about them, and even write radiology reports.

Google Health’s Med-Gemini-2D model demonstrates improved performance in writing chest X-ray radiology reports, exceeding previous state-of-the-art results by up to 12%.
🔬 They explore capabilities across 2D medical images, including X-rays, histopathology slides, and dermatology photos.
🔮 Med-Gemini-Polygenic is Google’s Health first language model capable of predicting health outcomes based on genomic data.

Importantly, their models are trained using free text reports that are freely available for most medical imaging, almost eliminating the need for costly expert labelling. We hope this will make it easier, faster, cheaper to train high-quality medical AI models in the future than has been possible in the past.

Google believe this research is a promising step forward in the field of medical AI – the ability to interpret complex multimodal medical data is a crucial step toward enabling high quality personalised care at scale. Of course, these are early results, and further research is required before safe clinical use – if you’re interested in joining us on this journey, please get in touch by filling in the form here: https://shorturl.at/aejpE

Full paper at https://arxiv.org/abs/2405.03162- we hope you enjoy it.

Enjoying these daily updates, subscribe below for more: Read Aloud For Me – Recommend AI Tools For Me

More on today’s OpenAI Announcement (ChatGPT-4o)

——————————————

New model. It’s called GPT-4o (terrible name) and has GPT-4-level intelligence. It’s 2x faster. 50% cheaper. 5x higher rate limits (compared to GPT-4-Turbo).
💡 What everyone is focused on: it’s cheaper! And faster! And with API access!
👀 What I am focused on: free users just got a mega performance boost which will likely reduce churn and the model architecture is a WINNER and worthy of attention – they went from three models to one; according to Andrej Karpathy, former OpenAI, they have built a “combined text-audio-vision model that processes all three modalities in one single neural network.”

——————————————

GPTs for all. Now every user, even free users, can access the “mini task bot” GPTs.
💡 What everyone is focused on: yay now everyone can use the GPT I built!
👀 What I am focused on: completely new user base, tens of millions new users testing and breaking capabilities

——————————————

More Voice. Now way more real-time (previously there was a 2-3 second lag). You can interrupt it mid-sentence. The voice assistant “picks up on emotion” (like fast breathing). I hate that phrase and would rather call it speech nuances. Fast multi-language translation. Performance improved for 50 languages (97% of the world’s population).
💡 What everyone is focused on: wow it sounds like Scarlett Johansson!
👀 What I am focused on: I’m already talking to ChatGPT Voice every morning. This is going to massively increase voice-first experiences. I think office spaces need to think about this asap. Think about the acoustics and EVERYONE talking to an AI assistant at once. It’s already an issue on my team.

——————————————

Vision on desktop. Now the desktop version can “see” your screen—only when you permission it to, not all the time. Sort of like generative AI alt text + chat. So you can ask it to describe a graph on your screen or presumably ask it questions about an article on your screen without a big lift.
💡 What everyone is focused on: lots of privacy concerns (I agree) and why do we need voice for code
👀 What I am focused on: HOLY MOLY THIS IS THE WINNING FEATURE. It’s basically a coworker on screen share with you 24/7, with no fatigue. I can imagine people working for hours straight with this on.

——————————————

Rollouts over the next few weeks.

If you like voice features (like talking to Siri but smarter), upgrade to Plus when it releases.

♻️ repost so others are in the know
🤖 follow Allie K. Miller to stay on top of AI

——————————————

Enjoying these daily updates, subscribe below for more: Read Aloud For Me – Recommend AI Tools For Me

A Daily chronicle of AI Innovations May 11th 2024: Weekly Summary – 🤖 DrEureka can automate robot training using LLMs
🚀 Free AI model rivals GPT-4 in language model evaluation
📰 X introduces Stories feature powered by Grok AI
🤖 Apple is developing its own AI chip for data center servers
🤝 Stack Overflow and OpenAI have announced an API partnership
🌟 Microsoft is developing a new AI language model
🖼️ OpenAI’s new tool detects 98% DALL-E 3 generated images
📣 Meta expands AI-powered creativity tools for advertisers
🎬 OpenAI’s ‘Media Manager’ will let creators opt out of AI training
🕵️‍♀️ Microsoft developed a secretive AI service for US spies
🧬 Goole DeepMicrosoft and Isomorphic Labs introduce AlphaFold 3
🧠 OpenAI’s Model Spec shares how it teaches its models to behave
🔍 Microsoft-LinkedIn study reveals rapid AI adoption in workplace & hiring
💬 Stability AI launches Stable Artisan, a Discord bot for image & video
🎵 ElevenLabs develops an AI model to generate song lyrics

DrEureka can automate robot training using LLMs

In robotics, one of the biggest challenges is transferring skills learned in simulation to real-world environments.  NVIDIA researchers have developed a groundbreaking algorithm called DrEureka that uses LLMs to automate the design of reward functions and domain randomization parameters—key components in the sim-to-real transfer process.

The algorithm works in three stages: first, it creates reward functions with built-in safety instructions; then, it runs simulations to determine the best range of physics parameters; finally, it generates domain randomization configurations based on the data gathered in the previous stages.

When tested on various robots, including quadrupeds and dexterous manipulators, DrEureka-trained policies outperformed those designed by human experts.

Source

Free AI model rivals GPT-4 in language model evaluation

Prometheus 2, a free and open-source language model developed by KAIST AI, has shown impressive capabilities in evaluating other language models, approaching the performance of commercial models like GPT-4.

The model was trained on a new pairwise comparison dataset called the “Preference Collection,” which includes over 1,000 evaluation criteria beyond basic characteristics. By combining two separate models – one for direct ratings and another for pairwise comparisons – the researchers achieved the best results.

In tests across eight datasets, Prometheus 2 showed the highest agreement with human judgments and commercial language models among all freely available rating models, significantly closing the gap with proprietary models.

Source

X introduces Stories feature powered by Grok AI

X (formerly Twitter) has launched a new feature, Stories, that provides AI-generated summaries of trending news on the platform. Powered by Elon Musk’s chatbot Grok, Stories offers Premium subscribers brief overviews of the most popular posts and conversations happening on X.

With Stories, users can quickly catch up on the day’s trending topics without having to scroll through countless posts. Grok generates these summaries based solely on the conversations happening on X about each news story rather than analyzing the original news articles themselves. While this approach is controversial, X believes it will pique users’ curiosity and potentially drive them deeper into the source material.

Source

Apple is developing its own AI chip for data center servers

Apple is developing its own AI chip for data center servers, known internally as Project ACDC (Apple Chips in Data Center). The chip will likely focus on running AI models (inference) rather than training them, which is where Nvidia currently dominates.

The company is working closely with TSMC (Taiwan Semiconductor Manufacturing Co) to design and produce these chips, although the timeline for launch is uncertain. With this move, the company aims to keep up with rivals like Microsoft and Meta, who have made significant investments in generative AI.

Source

Stack Overflow and OpenAI have announced an API partnership

OpenAI will use OverflowAPI to improve model performance and provide attribution to the Stack Overflow community within ChatGPT. Stack Overflow will use OpenAI models to develop OverflowAI and to maximize model performance.

The partnership aims to improve the user and developer experience on both platforms. The first set of integrations and capabilities will be available in the first half of 2024, and the partnership will enable Stack Overflow to reinvest in community-driven features.

Source

Microsoft is developing a new AI language model

Microsoft is developing a new, large-scale AI language model called MAI-1 to compete with Google and OpenAI. The model is overseen by Mustafa Suleyman, recently hired co-founder of Google DeepMind.

MAI-1 will be larger and more expensive than Microsoft’s previous smaller, open-source models, with roughly 500 billion parameters. Microsoft could preview the new model as soon as its Build developer conference later this month.

Source

OpenAI’s new tool detects 98% DALL-E 3 generated images

OpenAI has developed a new tool to detect if an image was created by DALL-E 3, its AI image generator. The tool can detect DALL-E 3 images with around 98% accuracy, even if the image has been cropped, compressed, or had its saturation changed. However, the tool is not as effective at detecting images generated by other AI models, only flagging 5-10% of images.

OpenAI’s new tool detects 98% DALL-E 3 generated images
OpenAI’s new tool detects 98% DALL-E 3 generated images

This image detection classifier is only available to a group of testers, including research labs and research-oriented journalism nonprofits through OpenAI’s Research Access Program.

OpenAI has also added watermarking to Voice Engine, its text-to-speech platform, which is currently in limited research preview.

Source

Meta expands AI-powered creativity tools for advertisers

Meta has expanded its generative AI tools for advertisers. Advertisers can request AI to generate entirely new images, including product variations in different colors, angles, and scenarios. The AI tools can add text overlays with different fonts, expand images to fit different aspect ratios like Reels and Feed, and generate ad headlines that match the brand’s voice.

The AI features will roll out globally to advertisers by the end of 2024.

Meta is also expanding its paid Meta Verified service for businesses to more countries. Different pricing tiers offer features like account support, profile enhancements, and better customer service access.

Source

OpenAI’s ‘Media Manager’ will let creators opt out of AI training

OpenAI is developing Media Manager, a tool that will enable creators and content owners to decide what they own and specify how they want their works to be included or excluded from machine learning research and training. This first-ever tool of its kind will help OpenAI identify copyrighted text, images, audio, and video across multiple sources and reflect creator preferences.

OpenAI aims to  have the tool in place by 2025 and set a standard across the AI industry with it.

Source

A Daily chronicle of AI Innovations May 10th 2024: 💥 OpenAI to unveil Google competitor on Monday 🎵 ElevenLabs unveils new AI music generator 🇺🇸 US lawmakers unveil bill to make it easier to restrict exports of AI models 📈 AI demand soars in the workplace 🤖 Stability AI introduces AI bot for Discord users 🎤 ElevenLabs debuts AI model for generating lyrics 💰 Anthropic CEO defends dual funding from Google and Amazon 🎥 Krea AI introduces video generation for paid subscribers 🎥 Synthflow launches AI voice agent teams for streamlined customer support

OpenAI to unveil Google competitor on Monday

OpenAI to unveil Google competitor on Monday
OpenAI to unveil Google competitor on Monday

  • OpenAI is reportedly set to launch an AI-powered search product on Monday, strategically timed just before Google’s annual I/O conference.
  • This new AI search tool from OpenAI may be integrated with its ChatGPT chatbot and feature enhancements such as internet search capabilities and possibly Bing integration, building on the foundation laid by previous versions of ChatGPT.
  • Amidst Google’s ongoing challenges, including an antitrust case and competition from other AI startups like Perplexity, OpenAI’s move is seen as a significant threat to Google’s longstanding dominance in the search industry.
  • Source

ElevenLabs unveils new AI music generator

    • ElevenLabs, a voice AI startup, has debuted a new model that creates songs from user prompts, leveraging its expertise in voice cloning and synthesis.
    • The company is showcasing this new capability by transforming user-generated ideas into music, akin to Microsoft’s strategy with its AI chatbot Copilot, and is exploring its application in areas like lullabies, jingles, and podcast intros.
    • While the AI music generation sector is growing, with ElevenLabs planning a marketplace for AI-generated songs, there are rising concerns about copyright infringement and the impact on artists’ jobs.

US lawmakers unveil bill to make it easier to restrict exports of AI models

  • US lawmakers introduced a bipartisan bill aimed at making it easier for the Biden administration to impose export controls on AI models to protect against foreign threats.
  • The proposed legislation would allow the Commerce Department to prevent Americans from collaborating with foreign entities on AI developments that could endanger U.S. national security.
  • The bill seeks to overcome legal obstacles that currently make it difficult to regulate the export of open source AI models, amidst concerns over their misuse for cyber attacks or creating biological weapons.

AI demand soars in the workplace

Microsoft and LinkedIn have published their ‘2024 Work Trend Index Annual Report’, revealing the rapid adoption of AI tools by employees, with 75% of knowledge workers using AI and nearly half starting within the last six months.

AI demand soars in the workplace
AI demand soars in the workplace

Here are the key points:

  • 78% of AI users are bringing their own AI tools to work, with 52% reluctant to use them for their important tasks. 
  • 66% of leaders say they wouldn’t hire someone without AI skills, and 71% prefer less experienced candidates with AI skills over more experienced ones without.
  • Power users who use AI extensively are reaping benefits in productivity, creativity, and job satisfaction.
  • Skills are projected to change by 68% by 2030, accelerated by generative AI.

Why does this matter?

The study serves as a wake-up call for organizations to move beyond experimentation and develop comprehensive strategies for AI implementation. As AI permeates all sectors, generations, and skill sets, early adopters will likely emerge as the leaders, while those hesitant to explore could risk falling behind. 

Source

Stability AI introduces AI bot for Discord users 

Stability AI has launched Stable Artisan, a multimodal-gen AI Discord bot that enables users to create images and videos using the Stable Diffusion 3 (SD3) and Stable Video Diffusion (SVD) models. 

https://youtu.be/MWfb30kWqTM?si=_TePwQX1A8xEj3hU

Stable Artisan incorporates several editing and customization features, including Search and Replace, Remove Background, Creative Upscale, Outpaint, Control Sketch, and Control Structure. The service is available through a paid subscription, with monthly plans ranging from $9 to $99, and a 3-day free trial. 

Stability AI is also working on a larger conversational chatbot called Stable Assistant, which will incorporate the company’s text-to-image and LLM technologies to assist users with various tasks through natural language conversations. While Stable Artisan currently does not include access to Stable Audio, Stable Code, or Stable LM, these features may be added in the future as the service continues to evolve.

Why does this matter?

Stable Artisan could empower creators lacking experience with complex AI models to generate high-quality content directly within their familiar Discord environment.

However, the paid subscription model could limit access, and the missing features hint at a future with a complete AI creative suite. 

https://stability.ai/news/stable-artisan

ElevenLabs debuts AI model for generating lyrics

ElevenLabs, a company that specializes in AI-powered voice cloning and synthesis, has revealed a new model that creates song lyrics based on user prompts. 

With this new model, ElevenLabs aims to impact the music industry by allowing users to generate custom lullabies, jingles, podcast intros, and potentially even popular songs. The company also plans to launch a marketplace where users can sell their AI-generated music.

While ElevenLabs has not yet shared details about the maximum length of songs the AI can generate, an example posted by the company’s Head of Design suggests that it will likely produce lyrics for a standard three-minute song. 

Why does this matter?

This AI tool has the potential to democratize songwriting, allowing even those without musical expertise to craft lyrics. This could be particularly impactful for budget-conscious creators or those with specific lyrical needs. However, it remains to be seen if it will integrate with composing melodies like Udio or Suno, which offer a more complete song creation experience.

Also, the AI lyrics’ originality and the tool’s training data on copyrighted music might raise uncertainties. 

Source

What Else Is Happening in AI on May 10th 2024

TikTok introduces “AI-generated” labels for third-party content

TikTok will automatically label AI-generated content on its platform and on third-party platforms, becoming the first social media platform to support Content Credentials metadata for AI transparency. (Link)

Anthropic CEO defends dual funding from Google and Amazon

Anthropic’s CEO says partnering with Google and Amazon ensures more independence than OpenAI’s Microsoft reliance. However, regulators are examining the impact on AI competition as Anthropic’s future training costs could reach $100 billion. (Link)

Krea AI introduces video generation for paid subscribers

Krea AI, a generative AI startup, has launched video generation capabilities for its highest-tier subscribers. The new feature allows users to create videos using a combination of key frame images and text prompts, with a timeline-based interface reminiscent of traditional video editing software. (Link

Synthflow launches AI voice agent teams for streamlined customer support 

Synthflow launches “Conversational AI Teams,” a feature that allows businesses to create multiple AI voice assistants to interact with customers and each other, all through a single phone number. These intelligent agents can handle tasks like scheduling, updating CRMs, and more, providing a seamless and efficient customer support experience. (Link)

AI TRAINING: Add shortcuts using ChatGPT’s Memory

AI TRAINING: Add shortcuts using ChatGPT’s Memory
AI TRAINING: Add shortcuts using ChatGPT’s Memory

A lesser-known feature of ChatGPT’s new Memory feature is that it can be programmed to store shortcuts, which can save you a lot of time in chat conversations when used effectively.

  • Go to ChatGPT’s website. You’ll need to have a Plus subscription to access Memory.
  • Click on your account name (bottom-left), select ‘Settings’, ‘Personalization’, and enable ‘Memory’.
  • Teach ChatGPT a list of shortcuts by adding: “Store in your memory that when I start a prompt using: ‘!si’ means that I want you to search on the Internet; ‘!ci’ means that I want you to create an image; and !da’ means that I want you to use your Advanced Data Analysis feature
  • That’s it! Try the shortcuts in your conversations and let us know how much time you save

AI RESEARCH: Llama-3 goes head-to-head with GPT-4

AI RESEARCH: Llama-3 goes head-to-head with GPT-4
AI RESEARCH: Llama-3 goes head-to-head with GPT-4

Lmsys just published a new deep dive into its Chatbot Arena data, comparing Meta’s open-source Llama 3 70B model against top models like GPT-4 — revealing surprising strengths in the open-source leader.

  • Llama-3 is the top open-source model on the Lmsys Leaderboard, featuring over 70,000 votes from users in Arena Battle testing.
  • Meta’s model shines in battles involving brainstorming and writing prompts, falling short in math, coding, and translation compared to top competitors.
  • Lmsys noted Llama-3’s tone is friendlier and more conversational than rivals, which the model exhibits in its victorious battles in the Arena.

Llama-3’s ability to compete with top-tier models in certain areas is a testament to the rapid progress of open-source — and that’s with Meta’s largest model still pending. The more granular comparison also provides useful details often lost in more general model benchmarking.

New AI Job Opportunities on May 10th 2024

A Daily chronicle of AI Innovations May 09th 2024: 💥 OpenAI to unveil Google competitor on Monday 🎵 ElevenLabs unveils new AI music generator 🇺🇸 US lawmakers unveil bill to make it easier to restrict exports of AI models  📈 AI demand soars in the workplace 🤖 Stability AI introduces AI bot for Discord users 🎤 ElevenLabs debuts AI model for generating lyrics

OpenAI to unveil Google competitor on Monday

OpenAI to unveil Google competitor on Monday
OpenAI to unveil Google competitor on Monday

  • OpenAI is reportedly set to launch an AI-powered search product on Monday, strategically timed just before Google’s annual I/O conference.
  • This new AI search tool from OpenAI may be integrated with its ChatGPT chatbot and feature enhancements such as internet search capabilities and possibly Bing integration, building on the foundation laid by previous versions of ChatGPT.
  • Amidst Google’s ongoing challenges, including an antitrust case and competition from other AI startups like Perplexity, OpenAI’s move is seen as a significant threat to Google’s longstanding dominance in the search industry.
  • Source

ElevenLabs unveils new AI music generator

    • ElevenLabs, a voice AI startup, has debuted a new model that creates songs from user prompts, leveraging its expertise in voice cloning and synthesis.
    • The company is showcasing this new capability by transforming user-generated ideas into music, akin to Microsoft’s strategy with its AI chatbot Copilot, and is exploring its application in areas like lullabies, jingles, and podcast intros.
    • While the AI music generation sector is growing, with ElevenLabs planning a marketplace for AI-generated songs, there are rising concerns about copyright infringement and the impact on artists’ jobs.

US lawmakers unveil bill to make it easier to restrict exports of AI models

  • US lawmakers introduced a bipartisan bill aimed at making it easier for the Biden administration to impose export controls on AI models to protect against foreign threats.
  • The proposed legislation would allow the Commerce Department to prevent Americans from collaborating with foreign entities on AI developments that could endanger U.S. national security.
  • The bill seeks to overcome legal obstacles that currently make it difficult to regulate the export of open source AI models, amidst concerns over their misuse for cyber attacks or creating biological weapons.

AI demand soars in the workplace

Microsoft and LinkedIn have published their ‘2024 Work Trend Index Annual Report’, revealing the rapid adoption of AI tools by employees, with 75% of knowledge workers using AI and nearly half starting within the last six months.

Here are the key points:

  • 78% of AI users are bringing their own AI tools to work, with 52% reluctant to use them for their important tasks. 
  • 66% of leaders say they wouldn’t hire someone without AI skills, and 71% prefer less experienced candidates with AI skills over more experienced ones without.
  • Power users who use AI extensively are reaping benefits in productivity, creativity, and job satisfaction.
  • Skills are projected to change by 68% by 2030, accelerated by generative AI.

Why does this matter?

The study serves as a wake-up call for organizations to move beyond experimentation and develop comprehensive strategies for AI implementation. As AI permeates all sectors, generations, and skill sets, early adopters will likely emerge as the leaders, while those hesitant to explore could risk falling behind. 

Source

Stability AI introduces AI bot for Discord users 

Stability AI has launched Stable Artisan, a multimodal-gen AI Discord bot that enables users to create images and videos using the Stable Diffusion 3 (SD3) and Stable Video Diffusion (SVD) models. 

https://youtu.be/MWfb30kWqTM?si=_TePwQX1A8xEj3hU

Stable Artisan incorporates several editing and customization features, including Search and Replace, Remove Background, Creative Upscale, Outpaint, Control Sketch, and Control Structure. The service is available through a paid subscription, with monthly plans ranging from $9 to $99, and a 3-day free trial. 

Stability AI is also working on a larger conversational chatbot called Stable Assistant, which will incorporate the company’s text-to-image and LLM technologies to assist users with various tasks through natural language conversations. While Stable Artisan currently does not include access to Stable Audio, Stable Code, or Stable LM, these features may be added in the future as the service continues to evolve.

Why does this matter?

Stable Artisan could empower creators lacking experience with complex AI models to generate high-quality content directly within their familiar Discord environment.

However, the paid subscription model could limit access, and the missing features hint at a future with a complete AI creative suite. 

https://stability.ai/news/stable-artisan

ElevenLabs debuts AI model for generating lyrics

ElevenLabs, a company that specializes in AI-powered voice cloning and synthesis, has revealed a new model that creates song lyrics based on user prompts. 

With this new model, ElevenLabs aims to impact the music industry by allowing users to generate custom lullabies, jingles, podcast intros, and potentially even popular songs. The company also plans to launch a marketplace where users can sell their AI-generated music.

While ElevenLabs has not yet shared details about the maximum length of songs the AI can generate, an example posted by the company’s Head of Design suggests that it will likely produce lyrics for a standard three-minute song. 

Why does this matter?

This AI tool has the potential to democratize songwriting, allowing even those without musical expertise to craft lyrics. This could be particularly impactful for budget-conscious creators or those with specific lyrical needs. However, it remains to be seen if it will integrate with composing melodies like Udio or Suno, which offer a more complete song creation experience.

Also, the AI lyrics’ originality and the tool’s training data on copyrighted music might raise uncertainties. 

Source

What Else Is Happening in AI on May 10th 2024

TikTok introduces “AI-generated” labels for third-party content

TikTok will automatically label AI-generated content on its platform and on third-party platforms, becoming the first social media platform to support Content Credentials metadata for AI transparency. (Link)

Anthropic CEO defends dual funding from Google and Amazon

Anthropic’s CEO says partnering with Google and Amazon ensures more independence than OpenAI’s Microsoft reliance. However, regulators are examining the impact on AI competition as Anthropic’s future training costs could reach $100 billion. (Link)

Krea AI introduces video generation for paid subscribers

Krea AI, a generative AI startup, has launched video generation capabilities for its highest-tier subscribers. The new feature allows users to create videos using a combination of key frame images and text prompts, with a timeline-based interface reminiscent of traditional video editing software. (Link

Synthflow launches AI voice agent teams for streamlined customer support 

Synthflow launches “Conversational AI Teams,” a feature that allows businesses to create multiple AI voice assistants to interact with customers and each other, all through a single phone number. These intelligent agents can handle tasks like scheduling, updating CRMs, and more, providing a seamless and efficient customer support experience. (Link)

AI TRAINING: Add shortcuts using ChatGPT’s Memory

AI TRAINING: Add shortcuts using ChatGPT’s Memory
AI TRAINING: Add shortcuts using ChatGPT’s Memory

A lesser-known feature of ChatGPT’s new Memory feature is that it can be programmed to store shortcuts, which can save you a lot of time in chat conversations when used effectively.

  • Go to ChatGPT’s website. You’ll need to have a Plus subscription to access Memory.
  • Click on your account name (bottom-left), select ‘Settings’, ‘Personalization’, and enable ‘Memory’.
  • Teach ChatGPT a list of shortcuts by adding: “Store in your memory that when I start a prompt using: ‘!si’ means that I want you to search on the Internet; ‘!ci’ means that I want you to create an image; and !da’ means that I want you to use your Advanced Data Analysis feature
  • That’s it! Try the shortcuts in your conversations and let us know how much time you save

AI RESEARCH: Llama-3 goes head-to-head with GPT-4

AI RESEARCH: Llama-3 goes head-to-head with GPT-4
AI RESEARCH: Llama-3 goes head-to-head with GPT-4

Lmsys just published a new deep dive into its Chatbot Arena data, comparing Meta’s open-source Llama 3 70B model against top models like GPT-4 — revealing surprising strengths in the open-source leader.

  • Llama-3 is the top open-source model on the Lmsys Leaderboard, featuring over 70,000 votes from users in Arena Battle testing.
  • Meta’s model shines in battles involving brainstorming and writing prompts, falling short in math, coding, and translation compared to top competitors.
  • Lmsys noted Llama-3’s tone is friendlier and more conversational than rivals, which the model exhibits in its victorious battles in the Arena.

Llama-3’s ability to compete with top-tier models in certain areas is a testament to the rapid progress of open-source — and that’s with Meta’s largest model still pending. The more granular comparison also provides useful details often lost in more general model benchmarking.

New AI Job Opportunities on May 10th 2024

A Daily chronicle of AI Innovations May 09th 2024: 🤖 OpenAI posts Model Spec revealing how it wants AI to behave 🧬 Google DeepMind unveils AlphaFold 3, the next generation of its protein prediction model 🧠 Neuralink faces setback as first human brain implant encounters problem 🕵️‍♀️  Microsoft developed a secretive AI service for US spies 🎨 Generate images on Midjourney Alpha 📝Copilot for Microsoft 365 to get auto-complete and rewrite to improve prompts 🏢New AI data center to be built at the failed Foxconn project site in Wisconsin 🤔Sam Altman says we are not taking AI’s impact on the economy seriously ✒️Typeface Arc replaces prompts; uses AI agent approach to ease marketing workflows 🎮Altera’s gaming AI agents get backed by Eric Schmidt, Former Google CEO

🤖 OpenAI posts Model Spec revealing how it wants AI to behave

OpenAI posts Model Spec revealing how it wants AI to behave
OpenAI posts Model Spec revealing how it wants AI to behave

OpenAI posts Model Spec revealing how it wants AI to behave
OpenAI posts Model Spec revealing how it wants AI to behave

 

  • OpenAI has introduced the first draft of Model Spec, a proposed framework aiming to shape how AI models respond, emphasizing assistance, humanity’s benefit, and adherence to social norms and laws.
  • The framework suggests specific rules for AI behavior, including compliance with laws, protection of privacy, and avoidance of NSFW content, with options to adjust settings like allowing NSFW content in certain contexts.
  • While the Model Spec seeks public feedback for future adjustments and doesn’t immediately affect existing models like GPT-4 or DALL-E 3, it’s envisioned as a living document to guide AI behavior improvement over time.
  • Source

🧬 Google DeepMind unveils AlphaFold 3, the next generation of its protein prediction model

  • Google DeepMind and Isomorphic Labs have released AlphaFold 3, a new AI model for predicting protein structures, including their interactions with various molecules such as DNA, RNA, and small molecules, thereby enhancing drug discovery possibilities.
  • This new version is more precise in mapping out complex groupings of molecules, significantly enhancing our ability to understand and predict molecular behavior compared to its earlier version.
  • Google will not open-source this version but has launched AlphaFold Server for non-commercial research use, aiming to balance intellectual property concerns with accessibility for scientific progress.
  • Source

🧠 Neuralink faces setback as first human brain implant encounters problem 

  • Neuralink admitted that some of the micro-thin threads from their N1 brain chip retracted after implantation in the first human patient, possibly due to air trapped in the skull during surgery, which affected the device’s data transmission rate.
  • Despite the retraction of several threads, Neuralink managed to increase the data transmission speed over time by optimizing their recording algorithm and improving signal translation into cursor movements.
  • The company is planning further implants, with goals to implant two more patients in the coming months and ten in total this year, while continuing to refine their technology and reporting developments to the FDA.
  • Source

🕵️‍♀️ Microsoft developed a secretive AI service for US spies

Microsoft has developed a top-secret generative AI model entirely disconnected from the internet so US intelligence agencies can safely harness the powerful technology to analyze top-secret info. The model based on GPT-4 is now live, answering questions, and will also write code.

Microsoft spent 18 months developing the model, which is “air-gapped” to ensure it is secure. This is the first time a model is fully isolated– meaning it’s not connected to the internet but is on a special network that’s only accessible by the U.S. government.

It can read and analyze files but cannot learn from them to stop sensitive information from entering the platform. It is yet to be tested and accredited by the intelligence agencies.

Why does this matter?

Intelligence agencies all over the world have been racing to be the first to harness generative AI. I guess we know who’s going to be the winner. If this AI tool is successful, it will fundamentally change the way intelligence agencies operate.

Source

What Else Is Happening in AI on May 09th 2024❗

📝Copilot for Microsoft 365 to get auto-complete and rewrite to improve prompts

In coming months, Microsoft Copilot will be updated with new features like auto-complete and ‘elaborate your prompt’ that offer suggestions to improve AI prompts. It aims to solve the problem of coming up with good prompts for generative AI. (Link)

🏢New AI data center to be built at the failed Foxconn project site in Wisconsin

President Joe Biden announced an AI data center to be built on the same site as the failed Foxconn project in Racine, Wisconsin. According to a White House press release, Microsoft is investing $3.3B in the project, creating up to 2,000 permanent jobs. (Link)

🤔Sam Altman says we are not taking AI’s impact on the economy seriously

At a Brooking’s Institute panel about AI and geopolitics on Tuesday, Altman said the discussions around AI’s effect on the economy–  like how it may lead to mass job replacement– died down this year compared to last. He said if we don’t take these concerns seriously enough going forward, it could be a massive issue. (Link)

✒️Typeface Arc replaces prompts; uses AI agent approach to ease marketing workflows

It is launching Typeface Arc technology, which enables a user to state a high-level marketing objective and then have the system automatically plan and generate all the assets, including emails, images, and notifications that are all connected. (Link)

🎮Altera’s gaming AI agents get backed by Eric Schmidt, Former Google CEO

Altera is the newest startup joining the fray to build a new guard of AI agents. It raised $9 million in an oversubscribed seed round, co-led by Eric Schmidt’s deep-tech fund, First Spark Ventures and Patron, the seed-stage fund co-founded by Riot Games alums. (Link)

AI TRAINING May 09th 2024

🎨 Generate images on Midjourney Alpha

Generate images on Midjourney Alpha
Generate images on Midjourney Alpha

Midjourney’s website is now accessible to anyone with more than 100 generated images, improving the experience when prompting images over its standard Discord group.

  1. Check that you’ve generated more than 100 images by typing /info in the Midjourney Discord group. If you have, head over to Midjourney Alpha.
  2. In the main menu, you can explore other creations and search prompts.
  3. Select where it says “imagine” and enter your prompt to generate an image.
  4. Add a reference image by selecting “+” or play with different parameters such as image size, stylization, or even weirdness by pressing the “slider control” button

AI RESEARCH on May 09th 2024

📶 AI usage surges in the workplace

AI usage surges in the workplace
AI usage surges in the workplace

Microsoft and LinkedIn just published their Work Trend Index Annual Report, revealing that AI adoption is surging in the workplace — calling 2024 the ‘year AI at work gets real’.

  • The report found that use of GenAI has doubled in the last six months, with 75% of knowledge workers using the tech in some capacity.
  • 78% of AI users are bringing their own AI to work — with 52% reporting they are reluctant to admit to its use.
  • 66% of leaders wouldn’t hire someone without AI skills, and 71% prefer less experienced candidates with AI aptitude over a more experienced one without it.
  • AI power users reported enhanced productivity, creativity, and job satisfaction compared to skeptical peers.

Why it matters: Employees are adopting AI at a rapid pace, regardless of if their own organizations are ready for the shift. As AI spreads across all sectors, generations, and skillsets, the early adopters are rising to the top — while those that aren’t at least exploring the tech are quickly running out of time

Trending AI Tools May 09th 2024

  • 📍GeoSpy – Uncover photo locations with AI
  • 🧑‍💻 LangChain – Connect LLMs to private data for context-aware applications
  • 📊 Abstra – Scale business processes with Python and AI
  • 🎨 Freepik Pikaso Upscaler – Integrated with Magnific, enlarge images without losing quality
  • 💬 Notion AI Q&A – Q&A is now open to the public, allowing users to ask and find information across their workspace
  • 🎵 Udio Audio Inpainting – Select a portion of an AI-generated music track and regenerate it

New AI Job Opportunities on May 09th 2024

  • 🎥 The Rundown – Video Content Creator
  • 🤖 Anthropic – Research Engineer, Human-Computer Interfaces
  • 👩‍💻 Adept AI – Solutions Engineer
  • 📝 Mistral AI – Data Annotation Technical Program Manager

A Daily chronicle of AI Innovations May 08th 2024: 👨‍⚕️ Google built medical AI that can train itself and outperform real doctors 🕵️‍♂️ Microsoft launches AI chatbot for spies 🖼️ OpenAI’s new tool detects 98% DALL-E 3 generated images 📣 Meta expands AI-powered creativity tools for advertisers 🎬 OpenAI’s ‘Media Manager’ will let creators opt out of AI training 🍎 Apple releases M4 chip at the ‘Let Loose’ event with powerful AI capabilities 📰 OpenAI strikes licensing deal with People magazine publisher 🤖 Amazon announces Bedrock Studio to simplify Gen AI app development 👨‍💻 Oracle introduces Code Assistant to accelerate enterprise software development 🚀 Red Hat launches RHEL AI and InstructLab to democratize enterprise AI 🍏 Apple unveils iPad AI upgrades

👨‍⚕️ Google built medical AI that can train itself and outperform real doctors

  • Google’s DeepMind has developed Med-Gemini, a self-training medical AI that outperforms doctors in accuracy, achieving a 91.1% score on medical benchmarks.
  • Med-Gemini uses an uncertainty-guided search functionality and has bested GPT-4 models and actual doctors on 10 out of 14 medical benchmarks.
  • Despite its capabilities, Google’s medical AI aims to assist rather than replace doctors, helping to alleviate their workload.
  • Source

🕵️‍♂️ Microsoft launches AI chatbot for spies

  • Microsoft has launched a GPT-4-based generative AI model tailored for US intelligence agencies that operates offline to analyze top-secret information securely.
  • The AI chatbot aims to facilitate secure conversations among spy agencies without the internet, addressing data breach and hacking concerns.
  • This initiative represents Microsoft’s first deployment of a major language model in a high-security environment, with a focus on mitigating the risks of data leaks while processing classified data
  • Source

🖼️ OpenAI’s new tool detects 98% DALL-E 3 generated images

OpenAI has developed a new tool to detect if an image was created by DALL-E 3, its AI image generator. The tool can detect DALL-E 3 images with around 98% accuracy, even if the image has been cropped, compressed, or had its saturation changed. However, the tool is not as effective at detecting images generated by other AI models, only flagging 5-10% of images.

OpenAI’s new tool detects 98% DALL-E 3 generated images

This image detection classifier is only available to a group of testers, including research labs and research-oriented journalism nonprofits through OpenAI’s Research Access Program.

OpenAI has also added watermarking to Voice Engine, its text-to-speech platform, which is currently in limited research preview.

Why does it matter?

Early experiences have shown that AI detectors don’t work. In fact, if they have high error rates, they could lead to false accusations. In 2023, OpenAI had to shut down its own AI detection software for text because of its poor accuracy.

So, if this detector is as good as OpenAI claims, we may be on the precipice of a revolutionary new capability to reliably detect AI-generated content, with huge implications across domains.

Source

📣 Meta expands AI-powered creativity tools for advertisers

Meta has expanded its generative AI tools for advertisers. Advertisers can request AI to generate entirely new images, including product variations in different colors, angles, and scenarios. The AI tools can add text overlays with different fonts, expand images to fit different aspect ratios like Reels and Feed, and generate ad headlines that match the brand’s voice.

Meta expands AI-powered creativity tools for advertisers
Meta expands AI-powered creativity tools for advertisers

The AI features will roll out globally to advertisers by the end of 2024.

Meta is also expanding its paid Meta Verified service for businesses to more countries. Different pricing tiers offer features like account support, profile enhancements, and better customer service access.

Why does it matter?

Integrating such powerful AI features could lead to more effective advertising campaigns and improved customer engagement with targeted marketing and personalized ads. However, it could also raise questions about transparency and potential misuse.

Source

🎬 OpenAI’s ‘Media Manager’ will let creators opt out of AI training

OpenAI is developing Media Manager, a tool that will enable creators and content owners to decide what they own and specify how they want their works to be included or excluded from machine learning research and training. This first-ever tool of its kind will help OpenAI identify copyrighted text, images, audio, and video across multiple sources and reflect creator preferences.

OpenAI aims to  have the tool in place by 2025 and set a standard across the AI industry with it.

Why does it matter?

Media Manager seems to be OpenAI’s response to growing criticism of its approach to developing AI models, which heavily scraps publicly available data from the web for training. Recently, 8 prominent U.S. newspapers sued OpenAI for copyright infringement.

On the other hand, OpenAI has formed mutually beneficial partnerships with platforms like Stack Overflow, Shutterstock, The Financial Times, and more to use their content.

So, OpenAI may be trying to meet creators in the middle, but if it is positioning itself as a fully ethical actor with this, we’ll take it with a grain of salt.

Source

🍏 Apple unveils iPad AI upgrades

Apple unveils iPad AI upgrades

Apple just revealed its new line of iPads at a company event in Cupertino, CA — featuring a custom M4 chip that enables advanced AI capabilities and a slew of new AI-powered features.

  • The bigger iPad Pro now features the M4 chip with an upgraded Neural Engine, which CEO Tim Cook calls “an outrageously powerful chip for AI”.
  • The M4 is capable of handling 38T operations per second, 4x the performance of previous models, allowing for the running of advanced AI.
  • New AI features on the Pro include a True Tone Flash for document scanning and new video, image, and music editing tools.
  • Prior to the event, a new report revealed that Apple is developing its own AI chips for data centers under the code name “Project ACDC”.

Source

What Else Is Happening in AI on May 08th 2024❗

🍎 Apple releases M4 chip at the ‘Let Loose’ event with powerful AI capabilities

Apple released its much-anticipated M4 chip at the “Let Loose” event. M4 is slated to spearhead Apple’s next generation of devices, with the iPad Pro leading the charge and powering the forthcoming OLED iPad Pro which is meticulously engineered to elevate the user experience to unprecedented heights. (Link)

📰 OpenAI strikes licensing deal with People magazine publisher

OpenAI has inked a licensing deal with Dotdash Meredith to bring the People magazine publisher’s content to ChatGPT and help train it’s AI models. Under the partnership, OpenAI will be able to display lifestyle and entertainment content in its chatbot from the many websites of one of the US’s largest digital and print publishers. (Link)

🤖 Amazon announces Bedrock Studio to simplify Gen AI app development 

Amazon is launching a new tool, Bedrock Studio, designed to let organizations experiment with generative AI models, collaborate on those models, and ultimately build generative AI-powered apps. Bedrock Studio is a “rapid prototyping environment” for generative AI. It also guides developers in evaluating, analyzing, fine-tuning, and sharing generative AI models. (Link)

👨‍💻 Oracle introduces Code Assitant to accelerate enterprise software development

Oracle has announced Code Assitant, an AI-powered service to help developers rapidly program apps based on Java, SQL, and the Oracle Cloud infrastructure. It will join tools like GitHub Copilot and Amazon CodeWhisperer to accelerate the app development lifecycle. However, Oracle hasn’t yet specified when this feature will be released. (Link)

🚀 Red Hat launches RHEL AI and InstructLab to democratize enterprise AI

At Red Hat Summit 2024, RedHat announced two major initiatives to bring the power of generative AI to the enterprise. Red Hat Enterprise Linux AI (RHEL AI), a foundation model platform for developing and running open-source language models, and InstructLab, a community project to enable domain experts to enhance AI models with their knowledge. (Link)

AI TRAINING: 🎥 Chat with YouTube videos using Gemini

Chat with YouTube videos using Gemini

Google Gemini’s new “Extensions” feature allows users to access external tools such as YouTube to chat with videos and get answers for free.

Step-by-step:

  1. Visit Google’s Gemini website. If Gemini is not available in your country, you’ll need to use a US-based VPN.
  2. Click on the gear icon located on the bottom-left, select Extensions, and turn on the YouTube one.
  3. Go back to the Chat interface and start your prompt using the following format: “@youtube Summarize the following video [Youtube URL]”

Pro tip: Try asking Gemini to explain advanced concepts discussed in a video, generating concrete examples, creating practice questions, and even asking for code snippets

Source

Trending AI Tools on May 08th 2024:

  • 🎨 Eraser AI – Technical design copilot that helps users edit documents and generate diagrams easily
  • 📊 Snaplet Seed – Cease manual data creation, get instant realistic data for your project
  • 🌐 Aria Browser AI – Opera’s browser AI can now create images and read out text responses
  • ⌛ Meaning – Block distracting apps with an AI screen time coach
  • 🤖 Dante Unlimited AI Messages – No more hidden fees for AI chatbot creation

New AI Job Opportunities May 08th 2024

A Daily chronicle of AI Innovations May 07th 2024: 🤖 Apple is developing its own AI chip for data center servers 🤝 Stack Overflow and OpenAI have announced an API partnership🌟 Microsoft is developing a new AI language model to rival GPT-4 and Google Gemini 📚 Study: Teachers can’t detect AI writing 🔮 Apple’s new M4 chip is focused on AI  🤖 Hugging Face has launched LeRobot, an open-source robotics toolkit 📸 Apple is testing a new “Clean Up” feature in its Photos app 🛡️ Google has launched Google Threat Intelligence 🇺🇸 US invests $285M in AI ‘Digital Twin’ technology  📡 Anduril Industries introduces Pulsar: AI modular electromagnetic warfare (EW) systems

🔮 Apple’s new M4 chip is focused on AI

  • Apple’s new M4 chip, designed for the latest OLED iPad Pro models, focuses on enhancing AI performance with capabilities like handling 38 trillion operations per second, which is significantly higher than its predecessors but still behind Qualcomm’s Snapdragon X Elite.
  • The M4 chip utilizes advanced second-generation 3nm technology, featuring a 10-core CPU with a mix of performance and efficiency cores, and supports technologies like mesh shading and ray tracing for improved graphical rendering.
  • Apple’s announcement of the M4 chip sets the stage for a comparison with Microsoft’s upcoming Arm-powered hardware and Windows on Arm enhancements, scheduled for release later this month, which aim to surpass Apple’s MacBook Air M3 in terms of CPU and AI performance.

Apple is developing its own AI chip for data center servers

Apple is developing its own AI chip for data center servers, known internally as Project ACDC (Apple Chips in Data Center). The chip will likely focus on running AI models (inference) rather than training them, which is where Nvidia currently dominates.

The company is working closely with TSMC (Taiwan Semiconductor Manufacturing Co) to design and produce these chips, although the timeline for launch is uncertain. With this move, the company aims to keep up with rivals like Microsoft and Meta, who have made significant investments in generative AI.

Why does it matter?

Apple has a long history of designing custom chips for its devices like iPhones, iPads, and Macs, which is probably what makes them stand out. Having custom AI chips could allow the tech giant more control over its “AI destiny” versus relying on suppliers like Nvidia.

Source

Stack Overflow and OpenAI have announced an API partnership

OpenAI will use OverflowAPI to improve model performance and provide attribution to the Stack Overflow community within ChatGPT. Stack Overflow will use OpenAI models to develop OverflowAI and to maximize model performance.

The partnership aims to improve the user and developer experience on both platforms. The first set of integrations and capabilities will be available in the first half of 2024, and the partnership will enable Stack Overflow to reinvest in community-driven features.

Why does this matter?

Stack Overflow partnered with Google Cloud to develop Overflow API and to give Google’s Gemini models access to its knowledge communities. Now it is forming a similar partnership with OpenAI. Despite concerns about copyright breaches, such partnerships seem to be trending where both the parties have much to gain, but it just reaffirms that the big AI players remain hungry for data.

Source

Microsoft is developing a new AI language model

Microsoft is developing a new, large-scale AI language model called MAI-1 to compete with Google and OpenAI. The model is overseen by Mustafa Suleyman, recently hired co-founder of Google DeepMind.

MAI-1 will be larger and more expensive than Microsoft’s previous smaller, open-source models, with roughly 500 billion parameters. Microsoft could preview the new model as soon as its Build developer conference later this month.

Why does this matter?

Microsoft’s development of MAI-1 shows that it is not entirely relying on it’s OpenAI investment to go big in AI. Now, it has entered the AI race truly, competing with state-of-the-art models from Google, Anthropic, even Meta’s Llama 400B which is in training, and OpenAI itself.

Source

What Else Is Happening in AI on May 07th 2024❗

🤖 Hugging Face has launched LeRobot, an open-source robotics toolkit

It is a comprehensive platform for developers, researchers, and hobbyists to train AI models, share data, and simulate environments, all while seamlessly integrating with various robotic hardware. The toolkit offers pre-trained models and integrates with physics simulators for testing without physical robots. Hugging Face is also collaborating with diverse partners to build the largest crowdsourced robotics dataset. (Link)

📸 Apple is testing a new “Clean Up” feature in its Photos app

By using gen AI for advanced image editing, this feature will allow you to effortlessly remove unwanted objects from your photos using a simple brush. Apple may preview this new feature during its upcoming “Let Loose” iPad event or at WWDC in June. (Link)

🛡️ Google has launched Google Threat Intelligence 

Google has launched Google Threat Intelligence
Google has launched Google Threat Intelligence

It is a combination of Mandiant’s expertise, VirusTotal’s community insights, and Google’s vast threat visibility. Google Threat Intelligence assists with external threat monitoring, attack surface management, digital risk protection, IoC analysis, and expertise. With Gemini, organizations can now quickly search through vast amounts of threat data to protect against cyber threats. (Link)

🇺🇸 US invests $285M in AI ‘Digital Twin’ technology 

The Biden administration is investing $285 million for a new “CHIPS Manufacturing USA institute” focused on digital twins for the semiconductor industry. This approach uses AI to create virtual chip replicas, accelerating the production of next-gen processors. Intel and Micron are also set to receive funding to boost the development of new processors. (Link)

📡 Anduril Industries introduces Pulsar: AI modular electromagnetic warfare (EW) systems 

Pulsar uses AI to quickly identify and counter current and future threats across the electromagnetic spectrum, including small and medium-size drones. With its integration of software-defined radio, GPUs, and diverse compute capabilities, Pulsar is changing how we defend against rapidly evolving threats in an increasingly complex battlefield. (Link)

AI TRAINING May 07th 2024

🎙️ Studio-quality audio in seconds with AI

Studio-quality audio in seconds with AI
Studio-quality audio in seconds with AI

Adobe’s AI-powered ‘Enhance Speech’ tool dramatically improves the quality of audio voice recordings with just a few clicks.

Step-by-step:

  1. Access Adobe’s Enhance Speech website. Sign in or sign up for free.
  2. Upload your audio by selecting “Choose file” or simply drag and drop the file.
  3. Wait a few seconds while Adobe’s AI analyzes and removes unwanted noises from your file.
  4. Download your studio-quality audio and hear the difference!

Pro tip: If you have a video file, you can extract the audio using free websites that extract audio from video and add the enhanced audio back to your video using free video editors like CapCut

AI RESEARCH on May 07th 2024

📚 Study: Teachers can’t detect AI writing

Study: Teachers can’t detect AI writing
Study: Teachers can’t detect AI writing

A series of studies from several German universities found that both novice and experienced teachers struggled to accurately distinguish between student-written and AI-generated texts.

The details:

  • The first study trialed pre-service teachers on AI and student text, correctly ID’ing only 45.1% of AI-generated submissions.
  • The second study tested experienced teachers, who correctly identified just 37.8% of AI-generated submissions.
  • The study revealed that teachers were overconfident in their ability to spot AI-generated text.
  • AI detection software also often falsely flags student-written texts as potentially AI-generated, which further complicates the issue.

Why it matters: AI’s writing capabilities are only getting better — and relying on teacher intuition or unreliable tools may be no more effective than guessing. Unless better tools become available, it may be time to pivot to enabling students to work with AI instead of penalizing them for it.

Trending AI Tools May 07th 2024

  • 📊 Osum – Perform deep market research in seconds: Try now (Use checkout code RUNDOWN for  25% off)*
  • 💫 Coachvox – The ultimate AI marketing tool for coaches with content
  • ✍️ Pressmaster – All-in-one platform simplifies press planning and distribution
  • 🧰 Lamini – Enterprises LLM platform that you can own
  • 🚀 Postman v11 – Supercharging development with AI and APIs
  • 🤖 Adobe Acrobat AI Assistant – Get quick answers and one-click summaries from PDFs
  • 🎨 Leonardo AI Content Reference – Enhance creative control for image generations

New AI Job Opportunities May 07th 2024

  • 🤝 The Rundown – Community Manager
  • 📈 Scale AI – Senior Analytics Engineer
  • ⚖ Anthropic – Risk Manager, Responsible Scaling Team
  • 💻 OpenAI – Tech Lead Manager – Supercomputing Scheduling

A Daily chronicle of AI Innovations May 06th 2024: 🤖 DrEureka can automate robot training using LLMs 🚀 Free AI model rivals GPT-4 in language model evaluation 📰 X introduces Stories feature powered by Grok AI 🔒 Privacy complaint filed against OpenAI 💰 JPMorgan launches an AI-powered tool for thematic investing ⏩ YouTube Premium introduces AI-powered “Jump ahead” feature 💊 AI is now set to transform the drug discovery industry 🎤 AI helps bring back Randy Travis’ voice in new song 💥 Jack Dorsey has left Bluesky, calls X a ‘freedom technology’ 🫠 Why Warren Buffet isn’t investing in AI🧠 Massive prompts can outperform fine-tuning for LLMs, researchers find

DrEureka can automate robot training using LLMs

In robotics, one of the biggest challenges is transferring skills learned in simulation to real-world environments.  NVIDIA researchers have developed a groundbreaking algorithm called DrEureka that uses LLMs to automate the design of reward functions and domain randomization parameters—key components in the sim-to-real transfer process.

The algorithm works in three stages: first, it creates reward functions with built-in safety instructions; then, it runs simulations to determine the best range of physics parameters; finally, it generates domain randomization configurations based on the data gathered in the previous stages.

When tested on various robots, including quadrupeds and dexterous manipulators, DrEureka-trained policies outperformed those designed by human experts.

Why does it matter?

DrEureka makes robot training accessible and cost-effective for businesses and researchers alike. We may witness increased adoption of robotics in industries that have previously been hesitant to invest in the technology due to the complexity and cost of training robots for real-world applications.

Source

Free AI model rivals GPT-4 in language model evaluation

Prometheus 2, a free and open-source language model developed by KAIST AI, has shown impressive capabilities in evaluating other language models, approaching the performance of commercial models like GPT-4.

The model was trained on a new pairwise comparison dataset called the “Preference Collection,” which includes over 1,000 evaluation criteria beyond basic characteristics. By combining two separate models – one for direct ratings and another for pairwise comparisons – the researchers achieved the best results.

Free AI model rivals GPT-4 in language model evaluation
Free AI model rivals GPT-4 in language model evaluation

In tests across eight datasets, Prometheus 2 showed the highest agreement with human judgments and commercial language models among all freely available rating models, significantly closing the gap with proprietary models.

Why does this matter?

By enabling user-defined evaluation criteria, Prometheus 2 can be tailored to assess language models based on specific preferences and real-life scenarios, opening up new possibilities for developing specialized AI applications across various domains. It’s also an opportunity to create niche models that are culturally sensitive and relevant.

Source

X introduces Stories feature powered by Grok AI

X (formerly Twitter) has launched a new feature, Stories, that provides AI-generated summaries of trending news on the platform. Powered by Elon Musk’s chatbot Grok, Stories offers Premium subscribers brief overviews of the most popular posts and conversations happening on X.

X introduces Stories feature powered by Grok AI
X introduces Stories feature powered by Grok AI

With Stories, users can quickly catch up on the day’s trending topics without having to scroll through countless posts. Grok generates these summaries based solely on the conversations happening on X about each news story rather than analyzing the original news articles themselves. While this approach is controversial, X believes it will pique users’ curiosity and potentially drive them deeper into the source material.

Why does this matter?

X’s Grok-powered Stories feature may reshape the way we consume news. As more platforms integrate AI news summarization tools, traditional media outlets may face challenges in maintaining reader engagement and revenue. However, the reliance on platform-specific conversations for generating summaries raises concerns about the potential spread of misinformation and the creation of echo chambers.

Source

🫠 Why Warren Buffet isn’t investing in AI

  • Warren Buffett predicts AI scamming will become a major growth industry, highlighting its capacity to generate realistic and misleading content to deceive individuals.
  • Buffett acknowledges AI’s potential for both beneficial and detrimental impacts, admitting his own lack of understanding of the technology.
  • Despite the hype around AI in the investment world, Buffett exercises caution, likening its unpredictable potential impact to that of nuclear weapons.
  • Source

🧠 Massive prompts can outperform fine-tuning for LLMs, researchers find

  • Researchers have discovered that using large prompts with numerous examples in LLMs can be more effective than traditional fine-tuning techniques.
  • “In-context learning” (ICL) becomes increasingly efficient with larger context windows, allowing for the effective use of hundreds to thousands of examples without the need for task-specific training.
  • The study indicates that while fine-tuning requires more initial data and resources, ICL with extensive prompts presents a viable and sometimes superior alternative, especially as models improve in processing long inputs.
  • Source

What Else Is Happening in AI on May 06th 2024❗

🔒 Privacy complaint filed against OpenAI

The maker of ChatGPT is facing a privacy complaint in the European Union (EU) for its “hallucination problem.” The complaint alleges violations of GDPR, including misinformation generation and lack of transparency on data sources. The report highlights concerns about accuracy, data access, and the inability of ChatGPT to correct incorrect information. (Link)

💰 JPMorgan launches an AI-powered tool for thematic investing

IndexGPT is a new range of thematic investment baskets created using OpenAI’s GPT-4 model. The tool generates keywords associated with a theme, which are then used to identify relevant companies through natural language processing of news articles. IndexGPT aims to improve the selection of stocks for thematic indexes, going beyond obvious choices and potentially enhancing trend-following strategies. (Link)

⏩ YouTube Premium introduces AI-powered “Jump ahead” feature

The AI-powered feature allows users to skip past commonly skipped sections of a video and jump to the next best point. It is currently available for the YouTube Android app in the US with English videos and can be enabled through the experiments page. (Link)

💊 AI is now set to transform the drug discovery industry

Generative AI is now rapidly generating novel molecules and proteins that humans may not have considered. AI models, such as Google’s AlphaFold, are accelerating the drug development process from years to months while increasing success rates. Experts predict that AI-designed drugs will become the norm in the near future, but they will still need to prove their efficacy in human trials. (Link)

🎤 AI helps bring back Randy Travis’ voice in new song

Country singer Randy Travis has released a new song, “Where That Came From,” his first since losing his voice to a stroke in 2013.

The vocals were created using AI software and a surrogate singer under the supervision of Travis and his producer. The result is a gentle tune that captures Travis’ relaxed style, reinforcing the potential of AI voice cloning in the right hands. (Link)

AI Tutorial on May 06th 2024:

🧠 How to use ChatGPT’s new ‘Memory’ feature

OpenAI has rolled out a new feature called “Memory” for ChatGPT plus users, enabling it to remember specific user details across chats.
Step-by-step:
  1. Head over to ChatGPT. You need a Pro account to enable this feature.
  2. Click on your account name (bottom-left), select Settings, Personalization, and enable Memory.
  3. Teach ChatGPT to remember specific details by explicitly stating them in the chat, e.g., “Remember that I am vegetarian when you recommend a recipe.”
  4. Manage your memories by clicking the “Manage” button located in the Memory section to view, edit, or delete specific memories.
That’s it! You can now have more personalized conversations across all your conversations 🎉

New AI Job Opportunities on May 06th 2024

A Daily chronicle of AI Innovations May 04th 2024: 🎥 World’s first AI-generated music video was just released ✈️ Air Force launches an AI-controlled F-16 fighter jet Microsoft ties executive pay to security following multiple failures and breaches 🤖 X is using Grok to publish AI-generated news summaries

🎥 World’s first AI-generated music video was just released

  • A pro music video created with OpenAI’s Sora, a text-to-video AI model, has been released by writer-director Paul Trillo, marking the first of its kind.
  • The video for “The Hardest Part” by Washed Out utilized 700 clips generated by Sora, with 55 selected for the final edit, showcasing the AI’s ability to transform written descriptions into video clips.
  • Despite not being publicly available yet, Sora’s early access was granted to a select group of artists, including Trillo, sparking interest and debate about the future of AI in creative processes.
  • Source

✈️ Air Force launches an AI-controlled F-16 fighter jet

  • An experimental AI-powered F-16 fighter jet, carrying Air Force Secretary Frank Kendall, performed advanced combat maneuvers, demonstrating significant advancements in military aviation technology.
  • The Air Force plans to have over 1,000 AI-enabled unmanned warplanes by 2028, with the AI-controlled F-16, known as Vista, marking a historic step in air combat by executing a dogfight against a human-piloted F-16.
  • Despite concerns over AI’s autonomous decision-making in warfare, the technology is seen as essential for future combat scenarios, aiming to reduce pilot risk and lower costs compared to traditional manned fighters.
  • Source

🤖 X is using Grok to publish AI-generated news summaries

  • X is using Grok to create AI-generated summaries for news and discussions trending on its platform, offering them under the feature “Stories on X” exclusively to premium subscribers.
  • Grok generates summaries based on users’ posts, covering a wide range of topics including news events and platform discussions, with a user highlighting its capability by showcasing stories on diverse subjects such as Apple’s earnings and debates on national debt.
  • Despite its innovative approach, Grok is noted for potential inaccuracies in its summaries, as evidenced by a past error where it misinterpreted a basketball term, highlighting the need for users to verify its outputs.
  • Source

AI Weekly Rundown April 26th to May 03rd 2024

🍎 iOS 18 may have OpenAI-powered gen AI Capabilities
🎥 China’s Vidu generates 16-second 1080P videos, matching OpenAI’s Sora
🤖 New S1 robot mimics human-like movements, speed, and precision
🚀 Gradient AI releases Llama-3 8B with 1M context
🤔 Mysterious “gpt2-chatbot” AI model bemuses experts
💻 GitHub’s Copilot Workspace turns ideas into AI-powered software
🏆 Amazon launches Amazon Q, the world’s most capable Gen AI assistant
🏥 Google’s Med-Gemini models outperform doctors
🕵️‍♂️ Apple has set up a secretive AI lab in Switzerland
📈 Better and faster LLMs via multi-token prediction: New research
📱 Anthropic launches an iOS app and a new plan for teams
💸 Google’s AI advancements urged Microsoft’s billion-$ OpenAI investment
🔍 Scale AI’s study finds popular LLMs overfit public benchmarks
🌍 Ukraine debuts the world’s first AI diplomat, Victoria Shi
🧠 Sam Altman is ready to spend $50 billion a year to build AGI

A Daily chronicle of AI Innovations May 03rd 2024: 👀 Bill Gates is still pulling the strings at Microsoft;  🥴 AI engineers report burnout as ‘rat race’ to stay competitive hits tech industry; 🇺🇦 Ukraine introduces AI spokesperson for state affairs; 📊 How much do LLMs overfit public benchmarks? 🔮 Sam Altman’s stance on the future of AI 🌟 New from #NVIDIAResearch, VILA is a vision language model that can 🤔 reason among multiple images, learn in context, and even understand videos.

How much do LLMs overfit public benchmarks? 

A new study by Scale AI raises concerns about the reliability of LLM benchmark tests. It uncovers LLM overfitting by evaluating them on a new (designed from scratch) dataset, GSM1k that mimics a popular benchmark, GSM8k.

How much do LLMs overfit public benchmarks? 
How much do LLMs overfit public benchmarks?

Key findings:

  • Overfitting: Many LLMs performed significantly worse on GSM1k compared to GSM8k, with some models dropping by as much as 13%. This suggests they’ve simply memorized the answers to benchmark problems rather than learning true reasoning skills.
  • Family Trends: Certain LLM families, particularly Mistral and Phi, showed consistent overfitting across different model sizes.
  • Frontier Models Shine: Newer, more advanced LLMs showed minimal signs of overfitting, suggesting they may be achieving genuine reasoning abilities.
  • Data Contamination Suspected: Analysis suggests data contamination from benchmark sets may be one factor contributing to overfitting.
  • Reasoning Still Present: Even overfitting models exhibited some capability to solve novel problems, although not at the level their benchmark scores suggested.

Overall, the study highlights the need for more robust and reliable methods for evaluating LLM reasoning abilities.

Why does it matter?

The dataset proves that overfitting may be creating major false impressions of model performance. As AI capabilities continue to advance, it is crucial to develop evaluation approaches that can keep pace and provide a more accurate picture of a model’s real-world potential.

Source

Sam Altman’s stance on the future of AI

During a recent appearance at Stanford University, Altman talked about the future of AI, calling GPT-4, a currently impressive AI model, to be the “dumbest model” compared to future iterations. According to Altman, the future is dominated by “intelligent agents,” AI companions that can not only follow instructions but also solve problems, brainstorm solutions, and even ask clarifying questions.

OpenAI isn’t just talking about the future, they’re actively building it. Their next-generation model, GPT-5, is rumored for a mid-2024 release and might boast video generation capabilities alongside text and image.

But the real moonshot is their active participation in developing AGI.

Despite the significant costs involved, Altman remains undeterred. He believes that the potential benefits, such as solving complex problems across various industries, outweigh the financial burden.

Watch the whole Q&A session here.

Why does this matter?

Altman’s bold comments on GPT-4 being the dumbest model suggest that OpenAI is aiming for something even grander, and GPT-5 could be a stepping stone toward it (the next-gen AI framework).

Source

👀 Bill Gates is still pulling the strings at Microsoft 

  • Bill Gates predicted the revolution in computing through AI agents in a memo before Microsoft’s partnership with OpenAI, and despite public appearances, has been heavily involved in guiding Microsoft’s AI strategy.
  • Under Gates’ influence, Microsoft developed Copilot, an AI tool based on OpenAI’s GPT, transforming the company into the world’s most valuable public entity by enhancing productivity in Microsoft products.
  • Gates has continued to play a pivotal role behind the scenes at Microsoft, advising on strategy, product development, and fostering key relationships, notably with OpenAI, even after stepping down from the board amid misconduct allegations.
  • Source

🥴 AI engineers report burnout as ‘rat race’ to stay competitive hits tech industry

  • AI engineers at Microsoft and other companies are reporting burnout due to the pressure to quickly release AI products, amidst an industry-wide “AI rat race.”
  • Microsoft’s focus on speed over ethics and safety in AI development has been criticized, even as the company forms a Responsible AI Council to address safety concerns.
  • An Amazon AI engineer’s account of being forced to work over a weekend, only for the project to be canceled, highlights a similar disregard for worker well-being and a preference for speed over accuracy and testing in AI product development.
  • Source

🇺🇦 Ukraine introduces AI spokesperson for state affairs

  • Ukraine’s Foreign Ministry introduced an artificial intelligence-powered spokesperson, Victoria Shi, to deliver official statements and interact with the press, marking a historic use of digital personnel for governmental communications.
  • Victoria Shi, created with the likeness and voice of Ukrainian singer Rosalie Nombre, will have her statements written and confirmed by humans, using AI for her visual presentation.
  • The ministry aims to utilize AI, including measures like QR codes for statement verification, to save time and money, particularly in the context of its ongoing conflict with Russia.
  • Source

🌟 New from #NVIDIAResearch, VILA is a vision language model that can 🤔 reason among multiple images, learn in context, and even understand videos.

Read our technical deep dive ➡️ https://nvda.ws/3QtMrsM. In the past, vision language models have struggled with in context learning and reasoning within multiple images.

No alternative text description for this image

What Else Is Happening in AI on May 03rd, 2024❗

🤖 OpenAI prepares to challenge Google with ChatGPT-powered search: OpenAI is building a search engine, search.chatgpt.com, potentially powered by Microsoft Bing. This leverages their existing web crawler and Bing’s custom GPT-4 for search, posing a serious threat to Google’s dominance. (Link)

🚫 Microsoft bans U.S. police use of Azure OpenAI for facial recognition

Microsoft has banned U.S. police from using Azure OpenAI Service for facial recognition, including integrations with OpenAI’s image-analyzing models. The move follows Axon’s controversial GPT-4-powered tool to summarize audio from the body camera. However, the ban has exceptions and doesn’t cover Microsoft’s other AI law enforcement contracts. (Link)

🌐 IBM expands AI and data software on AWS marketplace

IBM has significantly expanded its software offerings on the AWS Marketplace, making 44 products accessible to customers in 92 countries, up from just five. The move, part of a strategic collaboration with AWS, focuses on AI and data technologies like Watson x.data, Watson x.ai, and the upcoming Watson x.governance. (Link)

🔒 Google Cloud supports Azure and AWS; integrates AI for security

Google Cloud now supports Azure and AWS, enabling enterprises to manage security across multi-cloud platforms. AI integration with existing solutions streamlines user experience and addresses the security talent gap. The AI-powered design manages risks efficiently amid increasing cyber threats, while extensive support simplifies tasks for enterprises. (Link)

💸 Microsoft invests $2.2B in Malaysia’s cloud and AI transformation 

Microsoft is investing $2.2 billion over the next four years to support Malaysia’s digital transformation, its largest investment in the country’s 32-year history. The investment includes building cloud and AI infrastructure, creating AI skilling opportunities for 200,000 people, establishing a national AI Centre of Excellence, enhancing cybersecurity capabilities, and supporting the growth of Malaysia’s developer community. (Link)

Trending AI Tools on May 03rd 2024

  • 🚀 Waxaing AI – Project management tool for growth marketers
  • 🤖 Amazon Q – Generative AI-powered assistant from AWS
  • 🥁 Soundry AI – AI sound sample VST for music creation and DJing
  • 💬 NVIDIA ChatRTX – New features including multi-language voice support and more LLMs
  • 👩‍🎤 Synthesia Expressive AI Avatars – Create AI avatars with human-like expressions
  • 💻 Relume – Create a wireframe and UX copy in seconds with AI

New AI Jobs opportunity on May 03rd 2024

This week in AI – all the Major AI developments in a nutshell

  1. AI21 Labs introduced Jamba, a production-grade Mamba based model. By enhancing Mamba Structured State Space model (SSM) technology with elements of the traditional Transformer architecture, Jamba compensates for the inherent limitations of a pure SSM model. Jamba optimizes for memory, throughput, and performance—all at once.  It outperforms or matches other state-of-the-art models in its size class. Jamba has been released with open weights, licensed under Apache 2.0. Available on Hugging Face and coming soon to the NVIDIA API catalog [Details].
  2. Databricks introduced DBRX, an open, general-purpose LLM that uses a fine-grained mixture-of-experts (MoE) architecture with 132B total parameters of which 36B parameters are active on any input. Across a range of standard benchmarks, DBRX outperforms open LLMs like Mixtral, LLaMA2-70B and Grok-1. It surpasses GPT-3.5, and it is competitive with Gemini 1.0 Pro. It is an especially capable code model, surpassing specialized models like CodeLLaMA-70B on programming, in addition to its strength as a general-purpose LLM. The model is also available on the Databricks API [Details].
  3. Hume AI released Empathic Voice Interface (EVI), a first-of-its-kind conversational AI with emotional intelligence. EVI uses a new form of multimodal generative AI that integrates large language models (LLMs) with expression measures, which Hume refers to as an empathic large language model (eLLM). The eLLM enables EVI to adjust the words it uses and its tone of voice based on the context and the user’s emotional expressions [Demo | Details | wait list].
  4. Tencent introduced AniPortrait, a novel framework for generating high-quality animation driven by audio and a reference portrait image. Code and model weights have been released [Paper | GitHub]
  5. X announced an update to its AI chatbot Grok-1.5, with improvements in performance in coding and math-related tasks, and a context length of 128,000 tokens. Grok-1.5 will soon be available to early testers. Earlier Elon Musk had announced that all Premium subscribers on X will gain access to Grok this week, not just those on Premium+, as before [Details]
  6. Qwen (Alibaba Cloud) released Qwen1.5-MoE, a 2.7B, a small MoE model with only 2.7 billion activated parameters yet matching the performance of state-of-the-art 7B models like Mistral 7B and Qwen1.5-7B. Compared to Qwen1.5-7B, which contains 6.5 billion non-embedding parameters, it achieves a 75% decrease in training expenses and accelerates inference speed by a factor of 1.74 [Details].
  7. Claude 3 models dominates LMSYS Chatbot Arena Leaderboard. Claude 3 Opus tops the list beating GPT-4 Turbo, while Claude 3 Sonnet outperform older GPT-4 models and Claude 3 Haiku beating Mistral Large [Link].
  8. Adobe introduces structure reference feature for Firefly AI and GenStudio for brands. It enables taking one image and generating new ones that may be completely different stylistically, but whose internal elements are arranged and sized similarly to the first image [Details].
  9. Mata AI introduced OPT2I, a a training-free text-to-image (T2I) optimization-by-prompting framework that provides refined prompts for a T2I model that improve prompt-image consistency. The framework starts from a user prompt and iteratively generates revised prompts with the goal of maximizing a consistency score. OPT2I can boost the prompt-image consistency by up to 24.9% [Paper]
  10. OpenAI has started testing usage-based GPT earnings by partnering with a small group of US builders [Details].
  11. Adobe introduced Firefly Services and Custom Models. Firefly Services makes over 20 new generative and creative APIs available to developers. Custom Models, allows businesses to fine tune Firefly models based on their assets [Details].
  12. Nvidia announced a collaboration with Hippocratic AI , a healthcare company that is offering generative AI nurses, that range in specialties from “Colonoscopy Screening” to “Breast Cancer Care Manager,” and work for $9 an hour [Details].
  13. Worldcoin Foundation open-sourced the core components of its iris-scanning Orb’s software [Details].
  14. Emad Mostaque resigned from his role as CEO of Stability AI and from his position on the Board of Directors of the company to pursue decentralized AI [Details]
  15. Stability AI released Stable Code Instruct 3B, an instruction-tuned Code LM based on Stable Code 3B. With natural language prompting, this model can handle a variety of tasks such as code generation, math and other software development related queries [Details].
  16. Mistral AI released Mistral-7B-v0.2 Base model. This is the base model behind Mistral-7B-Instruct-v0.2 released in Dec, 2023 [Details]
  17. Open AI shared new examples of the Sora generations by visual artists, designers, creative directors and filmmakers [Details].

A Daily chronicle of AI Innovations May 02nd 2024: 👀 Sam Altman: “GPT-4 is the dumbest AI you’ll ever have to use” 🎵 TikTok and UMG strike new licensing deal including AI protections 📲 Anthropic’s ChatGPT rival Claude is now available on iOS 🤖 Atlassian launches Rovo, its new AI teammate 🤖 Better and faster LLMs via multi-token prediction: New research 💸 Google’s AI advancements urged Microsoft’s billion-dollar OpenAI investment 🤖 Sanctuary AI teams up with Microsoft to advance general-purpose robot AI 🗣️ Nvidia’s ChatRTX now supports voice queries and Google’s Gemma model 🤝 Atlassian launches Rovo: An AI assistant for enhanced teamwork 📊 MongoDB launches an AI app-building toolkit to help businesses use gen AI 🎨 Ideogram introduces Pro Tier: 12,000 fast AI image generations monthly

👀 Sam Altman: “GPT-4 is the dumbest AI you’ll ever have to use” 

  • OpenAI CEO Sam Altman described GPT-4 as the “dumbest” AI model users will have to interact with, anticipating significantly smarter versions in the future.
  • Altman emphasized that OpenAI’s AI models, including GPT-4, have considerable potential for improvement, with successive models like GPT-5 and GPT-6 expected to be significantly smarter.
  • Despite the high cost of development, which could range from $500 million to $50 billion annually, Altman is confident in the future value these advanced AI systems will bring to society.
  • Source

🎵 TikTok and UMG strike new licensing deal including AI protections

  • UMG has reached a “multi-dimensional” agreement with TikTok to bring back its artists, including Taylor Swift, Drake, and Olivia Rodrigo, to the platform’s over one billion users.
  • The deal includes provisions to collaborate on ensuring AI development in the music industry protects artists’ rights and addresses unauthorized AI-generated music on TikTok.
  • Additionally, the agreement will introduce new monetization opportunities for artists through TikTok’s expanding e-commerce capabilities and enhance tools for artist promotion, including analytics and integrated ticketing.
  • Source

📲 Anthropic’s ChatGPT rival Claude is now available on iOS 

  • Anthropic has launched an iOS app for its AI model Claude 3, which has been anticipated due to the company’s recent hiring of iOS engineers.
  • The app offers both free and premium access to Claude 3’s features, including its vision capabilities for real-time image analysis and summarization.
  • Anthropic also launched a new “Team” plan, priced at $30 per user per month, offers advanced administrative controls and collaboration features, aiming to secure a stronger foothold in the enterprise AI market.
  • Source

🤖 Atlassian launches Rovo, its new AI teammate 

  • Atlassian launched Rovo, an AI assistant designed to integrate with its products, making data from various tools easily accessible and enabling automation of workflows in applications like Jira and Confluence.
  • Rovo is based on Atlassian’s “cloud teamwork graph” and aims to simplify tasks for knowledge workers by assisting in finding, understanding, and taking action on work-related information.
  • The AI features Rovo Agents that can automate complex tasks and workflows, offering customization through natural language interface without the need for programming skills.
  • Source

Better and faster LLMs via multi-token prediction: New research

New research, apparently from Meta, has proposed a novel approach to training language models (LMs). It suggests that training LMs to predict multiple future tokens at once instead of predicting only the next token in a sequence results in higher sample efficiency. The architecture is simple, with no train time or memory overhead.

Better and faster LLMs via multi-token prediction: New research 
Better and faster LLMs via multi-token prediction: New research

 Figure: Overview of multi-token prediction

The research also provides experimental evidence that this training paradigm is increasingly useful for larger models and in particular, shows strong improvements for code tasks. Multi-token prediction also enables self-speculative decoding, making models up to 3 times faster at inference time across a wide range of batch sizes.

Why does it matter?

LLMs such as GPT and Llama rely on next-token prediction. Despite their recent impressive achievements, next-token prediction remains an inefficient way of acquiring language, world knowledge, and reasoning capabilities. It latches on local patterns and overlooks “hard” decisions.

Perhaps, multi-token prediction could bring a shift in how LMs learn. It could equip LLMs with deeper understanding and complex problem-solving capabilities. (or Meta just wasted their compute.)

Source

Anthropic launches an iOS app and a new plan for teams

Anthropic, the creator of the Claude 3 AI models, released a new iOS app named Claude. The app enables users to access AI models, chat with them, and analyze images by uploading them.

Anthropic also introduced a paid team plan, offering enhanced features like more chat queries and admin control for groups of five or more. The app is free for all users of Claude AI models, including free users, Claude Pro subscribers, and team plan members. The company will also roll out an Android version soon.

Why does it matter?

Though a little late with its mobile app, Anthropic has caught up with its competitors like OpenAI and Google, who have apps running for quite a while. The company decided to offer an app version because many users have been accessing its AI models through the web.

Source


Google’s AI advancements may have urged Microsoft’s billion-dollar OpenAI investment

Internal emails have revealed that Microsoft invested $1 billion in OpenAI in 2019 out of fear that Google was significantly ahead in its AI efforts.

Microsoft CTO Kevin Scott sent a lengthy email to CEO Satya Nadella and Bill Gates stating Google’s AI-powered “auto complete in Gmail” was getting “scarily good” and added that Microsoft was years behind in terms of ML scale.

The emails, with the subject line “Thoughts on OpenAI,”  were made public on Tuesday as part of the Department of Justice’s antitrust case against Google. A large section of Scott’s email was redacted. Check out the email here.

Why does it matter?

While some might call it paranoia, the well-timed move has undeniably paid off– the initial $1 billion has now turned into a multi-billion-dollar partnership with OpenAI.

While the email-surfacing highlights the growing scrutiny of competition in the tech industry, it also makes me wonder if Microsoft’s investment in OpenAI could have influenced the overall direction of AI research and development.

Source

What Else Is Happening in AI on May 02nd 2024❗

🤖 Sanctuary AI teams up with Microsoft to advance general-purpose robot AI

Sanctuary AI has announced a collaboration with Microsoft to develop AI models for general-purpose humanoid robots. The partnership will leverage Microsoft’s Azure cloud computing platform and AI technologies to enhance the capabilities of Sanctuary AI’s robots. (Link)

🗣️ Nvidia’s ChatRTX now supports voice queries and Google’s Gemma model

 Nvidia has updated its ChatRTX chatbot to support Google’s Gemma model, voice queries, and additional AI models. The chatbot, which runs locally on a PC, enables users to search personal documents and YouTube videos using various AI models, including ChatGLM3 and OpenAI’s CLIP model. (Link)

🤝 Atlassian launches Rovo: An AI assistant for enhanced teamwork

Atlassian has launched Rovo, an AI assistant designed to improve teamwork and productivity. Rovo integrates with Atlassian’s products and offers features such as AI-powered search, workflow automation, and integration with third-party tools like Google Drive, Microsoft SharePoint, and Slack. (Link)

📊 MongoDB launches an AI app-building toolkit to help businesses use gen AI

It has launched the MongoDB AI Applications Program, or MAAP, to help companies accelerate building and deployment of AI-powered applications. It brings consultancies and foundation models providers, cloud infrastructure, generative AI frameworks, and model hosting together with MongoDB Atlas to develop solutions for business problems. (Link)

🎨 Ideogram introduces Pro Tier: 12,000 fast AI image generations monthly

Ideogram has launched a paid Pro tier for its AI image generation platform, allowing users to generate up to 12,000 images per month at faster speeds. The platform utilizes AI algorithms to create high-quality images for various applications, including design, marketing, and content creation. (Link)

⚙️ Google Chrome gets Gemini shortcut

The details:
  • Users will be able to type ‘@gemini’ followed by a prompt in Chrome’s desktop address bar to get responses from the AI without navigating to a separate site.
  • The feature builds on existing shortcuts like ‘@bookmarks’ and ‘@history’, making AI chat as seamless as other browser functions.
  • Gemini’s Extensions feature now also connects the chatbot to Maps, YouTube, Hotels, and Workspace.
  • The features are only being rolled out to supported languages and countries.
Why it matters: Gemini just got a whole lot more accessible — with the shortcut and integrations not only boosting the chatbot’s reach, but also introducing a wave of non-AI users to the tech. Subtle but impactful changes like these are what drive serious shifts in user habits.

AI Training:

🎨 Create infinite styles with Midjourney

Midjourney’s new parameter feature called –sref random lets users generate images in completely random styles to help spark creativity.
Step-by-step:
  1. Visit Midjourney’s Discord. You will need a paid subscription to use this feature.
  2. Type -sref random after your prompt to grab a completely random style.
  3. To create a new image based on a previous style, grab the URL of the image, and type -sref [URL]
Example prompt: “Portrait of a woman smiling –sref https://www.…”

AI RESEARCH

💊 AI model predicts drug effectiveness

Image source: DALL-E 3
Ohio State University researchers just developed CURE, an AI model that can accurately estimate drug treatment effects and effectiveness without clinical trials.
The details:
  • The model is trained on de-identified health records of over 3M patients, allowing it to gain a deep understanding of patient characteristics.
  • CURE outperformed seven other leading AI models in treatment effectiveness predictions, with improvements of 7-8% on key benchmarks.
  • The AI predictions are closely aligned with clinical trial findings in tests, showcasing the potential to generate insights that streamline drug testing.
Why it matters: With the ability to crunch massive medical datasets, CURE represents a significant step towards systems that can reliably estimate real-world drug effectiveness — potentially accelerating the discovery of new treatments without the cost and long timeframes of traditional clinical trials.

Trending AI Tools

  • 📱 Claude iOS App – The powerful AI assistant by Anthropic, now on your iPhone
  • 🎵 Udio New Features – Generate AI music longer than 2 minutes and extend tracks up to 15 minutes
  • ✨ Loom AI workflow – Turn any loom video into share-ready docs in a click
  • 🧠 Atlassian Rovo – AI Teammate for data-driven business decisions
  • 📊 Kratful – AI-driven feedback analysis for product optimization
  • 🎨 Stability Matrix – Simplify Stable Diffusion UIs and models management

New AI Job Opportunities

  • 👩‍💻 Perplexity AI – Program Manager – Community
  • 🏙️ Metropolis – Senior Director of Sales Engineering
  • 💻 Luma AI – Senior Backend Software Engineer
  • ✍️ Cohere – Copywriter / Copy Editor (part-time, contractor)

A Daily chronicle of AI Innovations May 01st 2024: 🏆 Amazon has launched Amazon Q, a Gen AI assistant for businesses and developers
🏥 Google’s Med-Gemini models outperforms doctors 🕵️‍♂️ Apple has set up a secretive AI lab in Switzerland 💰 Google to pay News Corp $5-6 million per year to develop AI content and products 💬 Yelp is launching an AI chatbot to help consumers connect with relevant businesses 🍎 Apple is testing Safari 18 with new features: Intelligent Search and Web Eraser ⚖️ Eight US newspapers have sued Microsoft and OpenAI for copyright infringement 🏥 A study of 16000 patients reveals that AI ECG alert systems significantly lower all-cause mortality

Amazon has launched Amazon Q, a Gen AI assistant for businesses and developers

The details:
  • Q Developer offers industry-leading code generation, testing, debugging, reasoning, and agents for step-by-step planning.
  • Q Business connects to company data repositories, enabling users to easily get answers, summarize info, analyze trends, and interact with enterprise data.
  • A new Q Apps feature allows non-technical users to create custom AI applications using natural language prompts from company data.
  • AWS VP Dr. Swami Sivasubramanian said Q can help employees become ‘over 80% more productive’ at their jobs.

Amazon has launched Amazon Q, a generative AI assistant designed for developers and businesses. It comes in three distinct offerings:

  • Amazon Q Developer frees up precious time by handling tedious tasks like testing, debugging, and optimizing AWS resources so developers can focus on core coding and innovation.
  • Amazon Q Business connects to 40+ enterprise data sources and equips employees with a data-driven digital assistant to answer questions, create reports, and provide insights based on enterprise data repositories.
  • Amazon Q Apps allows non-technical employees to build generative AI applications using natural language prompts.

Amazon is driving real-world impact by offering a free tier for Q Developer and reporting early customer productivity gains of over 80%. Amazon Q Developer Pro is available for $19/user/month and Amazon Q Business Pro for $20/user/month. A free trial of both Pro tiers is available until June 30, 2024.

Why does it matter?

By introducing a free tier for Q Developer and the user-friendly nature of Q Apps, Amazon could accelerate innovation across the software development lifecycle and business workflows. This could empower domain experts and business leaders to use AI to solve their specific challenges directly, leading to more tailored AI applications across various industries.

Source

Google’s Med-Gemini models outperforms doctors

Researchers from Google and DeepMind have introduced Med-Gemini, a family of highly capable multimodal AI models specialized in medicine. Based on the strengths of the Gemini models, Med-Gemini shows significant improvements in clinical reasoning, multimodal understanding, and long-context understanding. Models can be customized to fit novel medical modalities through specialized encoders, and web searches can be used for up-to-date information.

Google’s Med-Gemini models outperforms doctors
Google’s Med-Gemini models outperforms doctors

Med-Gemini has shown state-of-the-art performance on 10 of 14 medical benchmarks, including text, multimodal, and long-context applications. Moreover, the models achieved 91.1% accuracy on the MedQA (USMLE) benchmark, exceeding the previous best models by 4.6%. Its strong performance in summarizing medical notes, generating clinical referral letters, and answering electronic health record questions confirms Med-Gemini’s potential real-world use.

Why does it matter?

These models can reduce the administrative burden on healthcare professionals by outperforming human experts in tasks like medical text summarization and referral letter generation. Moreover, Med-Gemini’s ability to engage in multimodal medical dialogues and explain its reasoning can lead to more personalized and transparent care, reduce misdiagnosis due to lack of physician knowledge, and save lives and money.

Source

Apple has set up a secretive AI lab in Switzerland

Since 2018, the company has quietly hired 36 AI experts from Google, including notable figures like Bengio and Ruoming Pang, for its secretive “Vision Lab.” The lab focuses on building advanced AI models and products, and it is particularly interested in text and visual-based AI systems akin to OpenAI’s ChatGPT. Apple has also acquired AI startups FaceShift and Fashwall, which are likely contributing to the establishment of the new lab.

Why does it matter?

Apple may have been fashionably late to AI development, but quietly setting up the Zurich lab and primary AI development centers in California and Seattle signifies the company’s AI ambitions.

Source

What Else Is Happening in AI in May 2024❗

💰 Google to pay News Corp $5-6 million per year to develop AI content and products

While News Corp denies any specific AI licensing deal, the arrangement highlights a growing trend of tech giants licensing news archives for language model training. Similar deals were inked between OpenAI and the Financial Times, showing the importance of quality data. (Link)

💬 Yelp is launching an AI chatbot to help consumers connect with relevant businesses

The chatbot uses OpenAI’s LLMs and Yelp’s data to understand user problems and provide relevant professional suggestions. Yelp also introduces a “Project Ideas” section for personalized recommendations and checklists. Meanwhile, restaurants are getting a revamped guest management system for better staff utilization, real-time table status, and customer updates. (Link)

🍎 Apple is testing Safari 18 with new features: Intelligent Search and Web Eraser

Intelligent Search uses Apple’s on-device AI to identify topics and key phrases for summarization. Web Eraser allows users to persistently remove unwanted content from web pages. Apple is also working on an AI Visual Lookup feature for 2025, allowing users to obtain product information from images. These AI enhancements will debut with iOS 18 and macOS 15 at WWDC in June. (Link)

⚖️ Eight US newspapers have sued Microsoft and OpenAI for copyright infringement

These newspapers, owned by Alden Global Capital’s MediaNews Group, allege that the companies misused their articles to train Copilot and ChatGPT without permission or payment. The New York Times, The Intercept, Raw Story, and AlterNet have filed similar lawsuits. The newspapers claim that the AI systems reproduce their content verbatim and generate fake articles that damage their reputation. (Link)

🏥 A study of 16000 patients reveals that AI ECG alert systems significantly lower all-cause mortality

The AI was trained on over 450,000 ECG tests and survival data to predict a patient’s risk of death. Physicians were alerted when a patient’s ECG indicated they were in the top 5% risk category. The AI reduced overall deaths by 17% and cardiac deaths by 93% for high-risk patients. (Link)

🍎 Apple poached 30+ Google experts to open a secret AI lab

  • Apple has reportedly opened a secret AI research lab in Zurich, known as the “Vision Lab,” after hiring at least 36 AI experts from Google.
  • The Zurich-based “Vision Lab,” led by former Google AI head John Giannandrea, has already produced significant research in generative AI, focusing on models that interpret text and imagery to deliver precise results.
  • Despite Apple’s silent approach in AI research, leading to perceptions of its lateness in the AI race, the company has been discreetly advancing cutting-edge AI technology and maintaining a low profile in recruitment and product development.

👽 Mysterious “gpt2-chatbot” AI model appears suddenly, confuses experts

  • A new chatbot named “gpt2-chatbot” has appeared on the LMSYS Chatbot Arena, sparking speculation that it might be a secret test of OpenAI’s upcoming models, such as GPT-4.5 or GPT-5, although its performance has not significantly surpassed that of existing models like GPT-4 Turbo.
  • Early user reports praise the mysterious model for its impressive reasoning and ability to answer challenging AI questions effectively, but detailed testing is limited due to a rate restriction of eight queries per day.
  • Despite ongoing speculation and hints by OpenAI’s CEO, the exact nature and capability of the “gpt2-chatbot” remain unclear, with some suggesting it could be an OpenAI preview.

🌐 Apple reportedly working on AI-enabled Safari

  • Apple is developing Safari 18 with features like a personalized user interface, “Web Eraser,” and AI-powered “Intelligent Search,” set to release alongside iOS 18 and macOS 15.
  • “Intelligent Search” in Safari 18 will use Apple’s Ajax language-learning model for on-device AI technology to summarize webpage content.
  • Safari 18’s “Web Eraser” feature will allow users to delete specific parts of web pages, enhancing privacy by remembering these changes even after closing the tab or window.

A Daily Chronicle of AI Innovations in April 2024

  • AI: The Ultimate Sherlocking?
    by /u/mintone (Artificial Intelligence) on July 26, 2024 at 12:16 pm

    submitted by /u/mintone [link] [comments]

  • Speech-to-Text Solution for Multilingual Sentences / Mixed-language speech
    by /u/simbaninja33 (Artificial Intelligence Gateway) on July 26, 2024 at 11:54 am

    I am looking for a speech-to-text solution, either paid or open-source, that can accurately transcribe speech containing a mix of two languages within the same sentence. I have explored options like Microsoft Azure, Google Cloud, and OpenAI, but haven't found a satisfactory solution yet. For example, I need the solution to handle sentences like: "I have tried the restaurant yesterday, it is muy muy bueno, they serve some of the pizza, que haria mi abuela super celoza de la receta." "I went to the store y compré un poco de pan because we were running low." I have already tried Microsoft Azure, which can handle multiple languages, but only when they are not mixed within the same sentence (as mentioned in their documentation). Google Cloud's speech-to-text fails to accurately transcribe mixed-language speech, and OpenAI doesn't seem to offer this functionality. I am open to both continuous real-time speech recognition and file-based recognition. For real-time applications, I am also willing to consider workarounds, such as implementing a "button" that can be clicked to quickly switch between the main language and the second language. If anyone has experience with a solution that can handle this type of mixed-language speech recognition, I would greatly appreciate any suggestions or recommendations. Thank you in advance for your help! submitted by /u/simbaninja33 [link] [comments]

  • Any open source AI model with web search abilities?
    by /u/david8840 (Artificial Intelligence Gateway) on July 26, 2024 at 11:45 am

    Is there any open source AI model with web search abilities? I want to be able to ask it questions which require real time internet searching, for example "What is the weather like now in NY?" submitted by /u/david8840 [link] [comments]

  • Which companies are leading the way in AI detection? (for audio/video deepfakes, etc.?)
    by /u/ProfessionalHat3555 (Artificial Intelligence Gateway) on July 26, 2024 at 11:21 am

    So I was listening to the most recent Bill Simmons pod w/ Derek Thompson where they discuss conspiracy theories and AI shit-detection (40:00-48:00 if you're curious)... 1ST Q: what companies are you aware of that are already working on AI detection? 2ND Q: where do you think the AI detection slice of the market is going? Will there be consumer-grade products that we can use to run, say, a political video through a detection software & get a % of realness rating on it? Will these tools ONLY be available to big conglomerates who become the purveyors of truth? 3RD Q: If we're UNABLE to do this at-scale yet, what would need to happen tech-wise for AI detection to become more accessible to more people? (disclaimer: I'm not a dev) submitted by /u/ProfessionalHat3555 [link] [comments]

  • AI can't take people's jobs if there's no people.
    by /u/baalzimon (Artificial Intelligence Gateway) on July 26, 2024 at 10:53 am

    Looks more and more likely that human populations will decline in the future. Maybe the workforce will just be AI robots rather than young people. PEW: The Experiences of U.S. Adults Who Don’t Have Children 57% of adults under 50 who say they’re unlikely to ever have kids say a major reason is they just don’t want to; 31% of those ages 50 and older without kids cite this as a reason they never had them https://www.pewresearch.org/social-trends/2024/07/25/the-experiences-of-u-s-adults-who-dont-have-children/ submitted by /u/baalzimon [link] [comments]

  • UK School Under Fire for Unlawful Facial-Recognition Use
    by /u/Think_Cat1101 (Artificial Intelligence Gateway) on July 26, 2024 at 10:43 am

    https://www.msn.com/en-us/news/technology/uk-school-under-fire-for-unlawful-facial-recognition-use/ar-BB1qEmeX?cvid=6dfe65854c6e4c2ad473b0e649e795b2&ei=10 submitted by /u/Think_Cat1101 [link] [comments]

  • OpenAI reveals 'SearchGPT'
    by /u/Mindful-AI (Artificial Intelligence Gateway) on July 26, 2024 at 10:41 am

    submitted by /u/Mindful-AI [link] [comments]

  • Amazon’s AI Chip Revolution: How They’re Ditching Nvidia’s High Prices and Speeding Ahead
    by /u/alyis4u (Artificial Intelligence Gateway) on July 26, 2024 at 9:23 am

    Six engineers tested a brand-new, secret server design on a Friday afternoon in Amazon.com’s chip lab in Austin, Texas. Amazon executive Rami Sinno said on Friday during a visit to the lab that the server was full of Amazon’s AI chips, which compete with Nvidia’s chips and are the market leader.https://theaiwired.com/amazons-ai-chip-revolution-how-theyre-ditching-nvidias-high-prices-and-speeding-ahead/ submitted by /u/alyis4u [link] [comments]

  • OpenAI's SearchGPT Is Coming For Google Search; Here Are The Features That Will Reportedly Make It Better
    by /u/vinaylovestotravel (Artificial Intelligence Gateway) on July 26, 2024 at 9:00 am

    Dubbed "SearchGPT," the tool will offer "fast and timely answers with clear and relevant sources" by referencing content from websites and news publishers, including OpenAI content partners such as News Corp (The Post's parent company) and The Atlantic. Read more: https://www.ibtimes.co.uk/openais-searchgpt-coming-google-search-here-are-features-that-will-reportedly-make-it-better-1725770 submitted by /u/vinaylovestotravel [link] [comments]

  • Deleting chats from Blackbox AI?
    by /u/Intelligent-Fig-7791 (Artificial Intelligence Gateway) on July 26, 2024 at 7:40 am

    How on earth do you delete chats from blackbox.ai ? it seems like all chats are public by default submitted by /u/Intelligent-Fig-7791 [link] [comments]

A Daily Chronicle of AI Innovations in April 2024

A daily chronicle of AI Innovations April 01st 2024

AI Innovations in April 2024.

Welcome to the April 2024 edition of the Daily Chronicle, your gateway to the latest Artificial Intelligence innovations! Join us as we uncover the most recent advancements, trends, and groundbreaking discoveries in the world of AI. Explore a realm where industry leaders gather at events like ‘AI Innovations at Work’ and where visionary forecasts shape the future of AI. Stay informed with daily updates as we navigate through the dynamic world of AI, uncovering its potential impact and exploring cutting-edge developments throughout this exciting month. Join us on this thrilling journey into the limitless possibilities of AI in April 2024.

Experience the transformative capabilities of AI with “Read Aloud For Me – AI Dashboard – AI Tools Catalog – AI Tools Recommender” – your ultimate AI Dashboard and Hub. Seamlessly access a comprehensive suite of top-tier AI tools within a single app, meticulously crafted to enhance your efficiency and streamline your digital interactions. Now available on the web at readaloudforme.com and across popular app platforms including Apple, Google, and Microsoft, “Read Aloud For Me – AI Dashboard” places the future of AI at your fingertips, blending convenience with cutting-edge innovation. Whether for professional endeavors, educational pursuits, or personal enrichment, our app serves as your portal to the forefront of AI technologies. Embrace the future today by downloading our app and revolutionize your engagement with AI tools.

AI Tools Catalog - AI Tools Recommender - Read ALoud For Me
AI Tools Catalog – AI Tools Recommender – Read ALoud For Me

A Daily chronicle of AI Innovations April 30th 2024: 🚀 Gradient AI releases Llama-3 8B with 1M context 🤔 Mysterious “gpt2-chatbot” AI model bemuses experts 💻 GitHub’s Copilot Workspace turns ideas into AI-powered software 📰 OpenAI collaborates with Financial Times to use its content in ChatGPT 🚀 Cohere’s Command R models family is accessible through Amazon Bedrock 📊 NIST launches a new platform for generative AI evaluation 🧬 ‘ChatGPT for CRISPR’ creates new genome-editing tools 💰 Microsoft to invest $1.7 billion in Indonesia’s AI and cloud infrastructure

Gradient AI releases Llama-3 8B with 1M context 

Gradient AI has released a new Llama-3 8B language model version called Llama-3-8B-Instruct-Gradient-1048k. This model’s key feature is its ability to handle extremely long context lengths up to 1 million tokens.

Gradient AI releases Llama-3 8B with 1M context 
Gradient AI releases Llama-3 8B with 1M context 

To extend the context window to 1 million tokens, Gradient AI used techniques like NTK-aware initialization of positional encodings, progressive training on increasing context lengths similar to prior work on long context modeling, and optimizations to train on huge GPU clusters efficiently. The model was trained on 1.4 billion tokens, a tiny fraction of Llama-3’s original pretraining data.

Why does it matter?

The 1M context window allows the Llama-3 8B model to process and generate text based on much larger inputs, like entire books or long documents. This could enable new applications in summarizing lengthy materials, answering questions that require referencing an extensive context and analyzing or writing on topics that require considering a large amount of background information.

Source

Mysterious “gpt2-chatbot” AI model bemuses experts

A mysterious new AI model called “gpt2-chatbot” is going viral. It was released without official documentation, and there is speculation that it could be OpenAI’s next model.

gpt2-chatbot shows incredible reasoning skills. It also gets difficult AI questions right with a more human-like tone.

Mysterious “gpt2-chatbot” AI model bemuses experts
Mysterious “gpt2-chatbot” AI model bemuses experts

On a math test, gpt2-chatbot solved an International Math Olympiad (IMO) problem in one try. This does not apply to all IMO problems, but it is still insanely impressive.

Mysterious “gpt2-chatbot” AI model bemuses experts
Mysterious “gpt2-chatbot” AI model bemuses experts

Also, many AI experts discuss the gpt2-chatbot’s better coding skills than the newest version, GPT-4 or Claude Opus. Without official documentation, we still don’t know who released it and for what purpose.

However, there are a couple of speculations going around in the industry that gpt2-chatbot is:

  • It’s secretly GPT-5 released early OpenAI can benchmark it
  • It’s OpenAI’s GPT-2 from 2019 finetuned with modern assistant datasets

You can try out gpt2-chatbot for free by visiting https://chat.lmsys.org direct chat. Unfortunately, with so many people trying it right now, there are slow response times and a maximum of 8 turns per conversation.

Why does it matter?

If the “gpt2-chatbot” model truly represents a major advancement in language generation and conversational abilities, it could accelerate the development of more advanced virtual assistants, chatbots, and other natural language processing applications. However, if the model’s capabilities are overstated or have significant limitations, it may lead to disappointment and a temporary setback in the progress of conversational AI.

Source

GitHub’s Copilot Workspace turns ideas into AI-powered software

GitHub is releasing a new AI-powered developer environment called Copilot Workspace. It allows developers to turn an idea into software code using natural language and provides AI assistance throughout the development process—planning the steps, writing the actual code, testing, debugging, etc.

The developer just needs to describe what they want in plain English, and Copilot Workspace will generate a step-by-step plan and the code itself. By automating repetitive tasks and providing step-by-step plans, Copilot Workspace aims to reduce developers’ cognitive strain and enable them to focus more on creativity and problem-solving. This new Copilot-native developer environment is designed for any device, making it accessible to developers anywhere.

Why does it matter?


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

Copilot Workspace could significantly lower the barrier to entry for those who can create software by automating much of the coding work. This could potentially enable a future with 1 billion developers on GitHub building software simply by describing what they want. Copilot Workspace could also make software development more accessible to non-technical people.

Source

What Else Is Happening in AI on April 30th 2034❗

📰 OpenAI collaborates with Financial Times to use its content in ChatGPT

The Financial Times has signed a deal with OpenAI to license its content for developing AI models and allow ChatGPT to answer queries with summaries attributable to the newspaper. It will help OpenAI enhance the ChatGPT chatbot with archived content from the FT, and the firms will work together to develop new AI products and features for FT readers. (Link)

🚀 Cohere’s Command R models family is accessible through Amazon Bedrock 

Amazon Bedrock developers can access Cohere’s Command R and Command R+ LLMs via APIs. This addition gives enterprise customers more LLM options, joining Claude 3 Sonnet, Haiku, Opus, Mistral 7B, Mixtral 8x7B, and Mistral Large. The Command R and R+ models are highly scalable, RAG-optimized, and multilingual across 10 languages. (Link)

📊 NIST launches a new platform for generative AI evaluation 

NIST announced the launch of NIST GenAI, a new program spearheaded to assess generative AI technologies, including text- and image-generating AI. NIST GenAI will release benchmarks, help create “content authenticity” detection (i.e., deepfake-checking) systems, and encourage software development to spot the source of fake or misleading AI-generated information. (Link)

🧬 ‘ChatGPT for CRISPR’ creates new genome-editing tools

ChatGPT has a specialized version called “GenomeGuide for CRISPR Research,” focusing on genetic engineering. It aims to assist researchers in designing new, more versatile gene-editing tools compared to the normal ones. It is also an AI assistant dedicated to genetic discoveries and provides R&D support in genetic engineering and CRISPR technology. (Link)

💰 Microsoft to invest $1.7 billion in Indonesia’s AI and cloud infrastructure 

Microsoft will invest $1.7 billion over the next 4 years in cloud and AI infrastructure in Indonesia, as well as AI skilling opportunities for 840,000 people and support for the nation’s growing developer community. These initiatives aim to achieve the Indonesian government’s Golden Indonesia 2045 Vision to transform the nation into a global economic powerhouse. (Link))

A Daily chronicle of AI Innovations April 29th 2024: ⚔️ China unveils Sora challenger 📰 OpenAI to train AI on Financial Times content 🫠 Meta’s AI ad platform is glitching and blowing through cash 🤖 Tesla and Baidu join forces to build cars that will drive themselves in China 💥 New iPad Pro may use an AI-enabled M4 chip 🍎 iOS 18 may have OpenAI-powered gen AI Capabilities 🎥 China’s Vidu generates 16-second 1080P videos, matching OpenAI’s Sora 🤖 New S1 robot mimics human-like movements, speed, and precision

🍎 iOS 18 may have OpenAI-powered gen AI capabilities

Apple has reportedly reinitiated talks with OpenAI to incorporate generative AI capabilities into the upcoming iOS 18 operating system, which will power the next generation of iPhones. The tech giant has been quietly exploring ways to enhance Siri and introduce new AI-powered features across its ecosystem. As of now, the companies are reportedly actively negotiating the terms of the agreement.

Apple is also in discussions with Google about licensing its Gemini chatbot technology. As of now, Apple hasn’t made a final decision on which partners it will work with, and there’s no guarantee that a deal will be finalized. The company may ultimately reach agreements with both OpenAI and Google or choose another provider entirely.

Why does this matter?

The renewed talks indicate Apple’s desperate attempt to accelerate its gen AI innovation and catch up with Big Tech. If successful, this collaboration would position Apple as a leader in AI-driven mobile devices, setting a new standard for chatbot-like interactions. Users can anticipate more sophisticated AI features, improved voice assistants, and a wider range of AI-powered applications on future iPhones.

Source

China’s Vidu generates 16-second 1080P videos, matching OpenAI’s Sora

At the ongoing Zhongguancun Forum in Beijing, Chinese tech firm ShengShu-AI and Tsinghua University have unveiled Vidu, a text-to-video AI model. Vidu is said to be the first Chinese AI model on par with OpenAI’s Sora, capable of generating 16-second 1080P video clips with a single click. The model is built on a self-developed visual transformation model architecture called Universal Vision Transformer (U-ViT), which integrates two text-to-video AI models: the Diffusion and the Transformer.

During a live demonstration, Vidu showcased its ability to simulate the real physical world, generating scenes with complex details that adhere to real physical laws, such as realistic light and shadow effects and intricate facial expressions. Vidu has a deep understanding of Chinese factors and can generate images of unique Chinese characters like pandas and loong (Chinese dragons).

Why does this matter?

Vidu’s launch represents a technical and strategic achievement for China. No other text-to-video AI model has yet been developed with cultural nuances with the intention of preserving national identity. Moreover, the integration of Diffusion and Transformer models in U-ViT architecture pushes the boundaries of realistic and dynamic video generation, potentially reshaping what’s possible in creative industries.

Source

New S1 robot mimics human-like movements, speed, and precision

Chinese robotics firm Astribot, a subsidiary of Stardust Intelligence, has previewed its advanced humanoid robot assistant, the S1. In a recently released video, the S1 shows remarkable agility, dexterity, and speed while doing various household tasks, marking a significant milestone in the development of humanoid robots.

Utilizing imitation learning, the S1 robot can execute intricate tasks at a pace matching adult humans. The video showcases the robot’s impressive capabilities, like smoothly pulling a tablecloth from beneath a stack of wine glasses, opening and pouring wine, delicately shaving a cucumber, flipping a sandwich, etc. Astribot claims that the S1 is currently undergoing rigorous testing and is slated for commercial release in 2024.

Why does this matter?

The AI-powered humanoid robot industry is booming with innovation and competition. OpenAI recently introduced two impressive bots: one for folding laundry with “soft-touch” skills and another for natural language reasoning. Boston Dynamics unveiled the Atlas robot, and UBTech from China introduced its speaking bot, Walker S. Now, Astribot’s S1 bot has amazed us with its incredible speed and precision in household tasks.

Source

⚔️ China unveils Sora challenger

  • China has developed a new text-to-video AI tool named Vidu, capable of generating 16-second videos in 1080p, akin to OpenAI’s Sora but with shorter video length capability.
  • The tool was created by Shengshu Technology in collaboration with Tsinghua University, and aims to advance China’s standing in the global generative AI market.
  • Vidu has been showcased with demo clips, such as a panda playing guitar and a puppy swimming, highlighting its imaginative capabilities and understanding of Chinese cultural elements.
  • Source

📰 OpenAI to train AI on Financial Times content

  • The Financial Times has made a deal with OpenAI to license their content and collaborate on developing AI tools, with plans to integrate FT content summaries, quotes, and links within ChatGPT responses.
  • OpenAI commits to developing new AI products with the Financial Times, which already utilizes OpenAI products, including a generative AI search function, indicating a deeper technological partnership.
  • This licensing agreement places the Financial Times among other news organizations engaging with AI, contrasting with some organizations like The New York Times, which is pursuing legal action against OpenAI for copyright infringement.
  • Source

🫠 Meta’s AI ad platform is glitching and blowing through cash

  • Meta’s automated ad platform, Advantage Plus shopping campaigns, has been heavily overspending and failing to deliver expected sales outcomes for advertisers.
  • Marketers have experienced unpredictable costs and poor performance with Advantage Plus, citing instances of ad budgets being rapidly depleted and a lack of transparent communication from Meta.
  • Despite efforts to address technical issues, ongoing problems with Advantage Plus have led some businesses to revert to manual ad buying and question the efficiency of AI-driven advertising on Meta’s platforms.
  • Source

🤖 Tesla and Baidu join forces to build cars that will drive themselves in China 

  • Elon Musk’s Tesla has partnered with Chinese tech giant Baidu to collect data on China’s public roads, aiming to develop and deploy Tesla’s full self-driving (FSD) system in China.
  • The partnership enables Tesla to meet local regulatory requirements by using Baidu’s mapping service, facilitating the legal operation of its FSD software on Chinese roads.
  • Elon Musk also claimed companies need to spend at least $10 billion on AI this year, similar to Tesla’s investment, to stay competitive.
  • Source

💥 New iPad Pro may use an AI-enabled M4 chip

  • The upcoming iPad Pro lineup is anticipated to feature the latest M4 chipset, marking a significant upgrade from the current M2 chipset-equipped models.
  • The new M4 chipset in the iPad Pro is expected to introduce advanced AI capabilities, positioning the device as Apple’s first truly AI-powered product.
  • Apple’s “Let Loose” event, scheduled for May 7, will also showcase new OLED iPad Pro variants and the first 12.9-inch iPad Air, alongside potential launches of a new Magic Keyboard and Apple Pencil.
  • Source

👁️ OpenAI hit with GDPR complaint over ChatGPT’s ‘hallucination’ failure

  • OpenAI faces a new GDPR privacy complaint in the EU due to ChatGPT’s generation of incorrect personal information without a means to correct it.
  • The complaint challenges ChatGPT’s compliance with the GDPR, emphasizing the right of EU citizens to have erroneous data corrected and OpenAI’s refusal to amend incorrect information.
  • OpenAI’s situation highlights tension with GDPR requirements, including rights to rectification and transparency, as authorities in various EU countries investigate or consider actions against the company.
  • Source

What Else Is Happening in AI on April 29th 2024❗

Ace the Microsoft Azure Fundamentals AZ-900 Certification Exam: Pass the Azure Fundamentals Exam with Ease

💄 Estée Lauder and Microsoft’s collaboration for beauty brands

Estée Lauder Companies (ELC) and Microsoft have launched the AI Innovation Lab to help ELC’s brands leverage generative AI. The collaboration aims to enable faster responses to social trends and consumer demands, as well as accelerate product innovation. (Link)

🚀 Oracle boosts Fusion Cloud apps with 50+ generative AI capabilities

Oracle has launched new generative AI features across its Fusion Cloud CX suite to help sales, marketing, and service agents automate and accelerate critical workflows. The AI capabilities will enable contextually-aware responses, optimized schedules for on-field service agents, targeted content creation, and AI-based look-alike modeling for contacts. (Link)

💬 Google’s new AI feature helps users practice English conversations

The chatbot, currently available in select countries through Search Labs or Google Translate on Android, provides feedback and helps users find the best words and conjugations within the context of a conversation. (Link)

🧠 OpenAI enhances ChatGPT with user-specific memory update

The update enables ChatGPT to provide more personalized and contextually relevant responses over time by storing details about users’ preferences and interactions. Users have control over the memory feature, including the ability to toggle it on or off, inspect stored information, and delete specific data entries. (Link)

🤝 Tech CEOs join DHS advisory board on AI safety and security

The US DHS has announced a blue-ribbon board that includes CEOs of major tech companies to advise the government on the role of AI in critical infrastructure. They will develop recommendations to prevent and prepare for AI-related disruptions to critical services that impact national economic security, public health, or safety. (Link)

A Daily chronicle of AI Innovations April 27th 2024:💥 Apple in talks with OpenAI to build chatbot 🧠 China developed its very own Neuralink 🚗 Tesla Autopilot has ‘critical safety gap’ linked to hundreds of collisions 🐱  🤖 Google’s AI bot can now help you learn English.

💥 Apple in talks with OpenAI to build chatbot 

  • Apple is in talks with OpenAI and Google to incorporate their AI technology into the iPhone’s upcoming features, aiming to debut new generative AI functionalities at the Worldwide Developers Conference.
  • Apple has struggled to develop a competitive AI chatbot internally, leading to the cancellation of some projects to refocus on generative AI technologies using external partnerships.
  • Choosing to partner with OpenAI or Google could mitigate past challenges with AI implementations, but also increase Apple’s dependency on these competitors for AI advancements.
  • LINK

🤖 Google’s AI bot can now help you learn English

  • Google has introduced a ‘Speaking practice’ feature for learning English, allowing users to converse with an AI bot on their phones, a tool that began offering feedback on spoken sentences in October 2023 and now supports continuous dialogues.
  • The ‘Speaking practice’ feature is available to Search Labs users in countries like Argentina, India, Mexico, Colombia, Venezuela, and Indonesia and may appear when translating to or from English on an Android device.
  • Unlike structured curriculum apps such as Duolingo, Babbel, and Pimsleur, Google’s approach allows for practicing English within conversational contexts, with the company expanding its AI’s language comprehension.
  • Source

AI Weekly Rundown April 2024 Week 4 [April 21 – April 28]

🍎 iOS 18 to have AI features with on-device processing
🧠 Many-shot ICL is a breakthrough in improving LLM performance
⚡ Groq shatters AI inference speed record with 800 tokens/second on LLaMA 3
📱 Microsoft launches its smallest AI model that can fit on your phone
🎨 Adobe survey says 50% of Americans use generative AI every day
👨‍💻 Microsoft hired former Meta VP of infrastructure
🖼️ Firefly 3: Adobe’s best AI image generation model to date
👓 Meta finally rolls out multimodal AI capabilities for its smart glasses
🧬 Profulent’s OpenCRISPR-1 can edit the human genome
🤖 NVIDIA acquires Run:ai; integrates it with DGX Cloud AI Platform
❄️ Snowflake enters the generative AI arena with Arctic LLM
⏳ Monetizing generative AI to take time, says Zuckerberg
🏭 Sanctuary AI launches Phoenix 7 robot for industrial automation
🛑 AI integration hits roadblocks for CIOs
💊 Moderna and OpenAI partner to accelerate drug development

If you are looking for an all-in-one solution to help you prepare for the AWS Cloud Practitioner Certification Exam, look no further than this AWS Cloud Practitioner CCP CLF-C02 book

🧠 China developed its very own Neuralink

  • Beijing Xinzhida Neurotechnology, backed by the Chinese Community Party, unveiled a brain-computer interface named Neucyber, currently successful in controlling a robotic arm via a monkey.
  • Neucyber, regarded as a competitive response to Neuralink, highlights the intensifying global race in developing brain-computer interfaces, though it has not advanced to human trials yet.
  • The long-term implications of such technology remain uncertain, stirring both intrigue and concern in the context of its potential impact on health and the broader tech industry.
  • Source

🚗 Tesla Autopilot has ‘critical safety gap’ linked to hundreds of collisions 

  • The National Highway Traffic Safety Administration (NHTSA) reported that Tesla’s Autopilot contributed to at least 467 collisions, including 13 fatalities, due to a “critical safety gap” in its design.
  • NHTSA criticized the Autopilot system for inadequate driver monitoring and staying active even when drivers are not paying sufficient attention, leading to “foreseeable misuse and avoidable crashes.”
  • The agency is also investigating the effectiveness of a software update issued by Tesla intended to improve the Autopilot’s driver monitoring capabilities, following continued reports of related crashes.
  • Source

A Daily chronicle of AI Innovations April 26th 2024: 💰 Elon Musk raises $6B to compete with OpenAI 🤖 Sanctuary AI unveils next-gen robots; 💻 CIOs go big on AI! 🧬 Moderna and OpenAI partner to accelerate drug development 📱 Samsung and Google tease collaborative AI features for Android 🧠 Salesforce launches Einstein Copilot with advanced reasoning and actions 📋 AuditBoard integrates AI-powered descriptions to cut audit busywork 🚫 LA Metro to install AI cameras on buses to issue tickets to illegal parkers 💊 EPFL and Yale researchers develop Meditron, a medical AI model 

Sanctuary AI unveils next-gen robots

Sanctuary AI, a company developing human-like intelligence in robots, unveiled its latest robot – Phoenix Gen 7. This comes less than a year after their previous generation robot.

The new robot boasts significant improvements in both hardware and software. It can now perform complex tasks for longer durations, learn new tasks 50 times faster than before, and have a wider range of motion with improved dexterity. The company believes this is a major step towards achieving human-like general-purpose AI in robots.

Why does it matter?

While Boston Dynamics headlines focus on robotic feats, Sanctuary AI’s progress could set a new standard for the future of work and automation. As robots become more human-like in their capabilities, they can take on complex tasks in manufacturing, healthcare, and other sectors, reducing the need for human intervention in potentially dangerous or repetitive jobs.

Source

CIOs go big on AI!

A new Lenovo survey shows that CIOs are prioritizing integrating AI into their businesses alongside cybersecurity.

However, there are challenges hindering rapid AI adoption, such as:

  • Large portions of organizations are not prepared to integrate AI swiftly (e.g., new product lines, supply chain).
  • Security concerns around data privacy, attack vulnerability, and ethical AI use.
  • Talent shortage in machine learning, data science, and AI integration.
  • Difficulty demonstrating ROI of AI projects.
  • Resource constraints – focusing on AI may take away from sustainability efforts.

Despite the challenges, there is still a positive outlook on AI:

  • 80% of CIOs believe AI will significantly impact their businesses.
  • 96% of CIOs plan to increase their investments in AI.

Why does it matter?

This highlights a significant transition where CIOs are now focused on driving business outcomes rather than just operational maintenance. As AI plays a crucial role, addressing the barriers to adoption will have far-reaching implications across industries seeking to leverage AI for competition, innovation, and efficiency gains. Overcoming the skills gap and security risks and demonstrating clear ROI will be key to AI’s proliferation.

Source

Moderna and OpenAI partner to accelerate drug development

Biotech giant Moderna has expanded its partnership with Open AI to deploy ChatGPT enterprise to every corner of its business. The aim is to leverage AI to accelerate the development of new life-saving treatments.

Here’s the gist: 

  • Moderna plans to launch up to 15 new mRNA products in 5 years, including vaccines and cancer treatments.
  • Their custom “Dose ID” GPT helps select optimal vaccine doses for clinical trials.
  • Moderna saw the creation of 750+ custom GPTs with 120 ChatGPT conversations per user per week.
  • The redesign aims for a lean 3,000-employee team to perform like 100,000 with AI force multiplication.

Why does it matter?

If Moderna can pull this off, it could mean a future where new life-saving drugs are developed at lightning speed. And who knows, maybe your next doctor’s visit will involve a friendly chat with a healthcare AI. Just don’t ask it to diagnose you on WebMD first.

Source

What Else Is Happening in AI on April 26th 2024❗

📱 Samsung and Google tease collaborative AI features for Android: Samsung and Google are teasing new AI features developed through their strong partnership. Recent social media posts from Samsung Mobile and Google’s Rick Osterloh confirm the companies are working together on AI and exploring opportunities. The collaboration aims to deliver the best Android ecosystem of products and services. (Link)

🧠 Salesforce launches Einstein Copilot with advanced reasoning and actions

Salesforce announced the general availability of its generative AI platform, Einstein Copilot, with new features like Copilot Actions and Analytics. Actions enable sales teams to optimize workflows and close more deals, while Analytics provides insights into Copilot’s usage and performance. Salesforce is also working on improving efficiency with smaller AI models. (Link)

📋 AuditBoard integrates AI-powered descriptions to cut audit busywork

AuditBoard, a cloud-based audit software company, has launched AuditBoard AI, an advanced AI feature to automate risk assessment descriptions. The AI-powered tool generates descriptions for risks and controls, reducing the time auditors spend on repetitive tasks and increasing efficiency. (Link)

🚫 LA Metro to install AI cameras on buses to issue tickets to illegal parkers

LA Metro equips buses with AI cameras to catch and ticket vehicles blocking bus lanes, aiming to improve bus times and accessibility. Violations will be human-reviewed before ticketing. The program, launching this year, could lead to AI-assisted traffic management in the future. (Link)

💊 EPFL and Yale researchers develop Meditron, a medical AI model 

Researchers from EPFL and Yale have developed Meditron, an open-source suite of medical AI models based on Meta’s Llama. Designed for low-resource settings, Meditron assists with clinical decision-making and diagnosis. The models, fine-tuned on high-quality medical data with expert input, have been downloaded over 30,000 times. (Link)

💰 Elon Musk raises $6B to compete with OpenAI

  • xAI, Elon Musk’s AI company, is nearing a funding round of $6 billion at a pre-money valuation of $18 billion, aiming to be a competitor to OpenAI.
  • The funding round has attracted significant interest from investors, including Sequoia Capital and Future Ventures, and terms were adjusted from an initial $3 billion at a $15 billion valuation due to demand.
  • X, Musk’s social network, not only has a stake in xAI but also integrates its chatbot Grok, showcasing xAI’s broader ambition to merge the digital with the physical through data from Musk’s companies.
  • Source

A Daily chronicle of AI Innovations April 25th 2024: 🤖NVIDIA acquires Run:ai, integrates it with DGX Cloud AI Platform ❄️Snowflake enters the generative AI arena with Arctic LLM ⏳Monetizing generative AI to take time, says Zuckerberg 🎥 Adobe unveils VideoGigagan: AI project upscaling blurry videos to HD 📱 OpenELM: Apple’s evolving AI strategy for iPhones ☁️ IBM acquires HashiCorp for $6.4 Billion to boost cloud business 🙂 Synthesia Introduces Emotions to AI Video Avatars 🤖 HubSpot introduces cutting-edge AI tools for SMBs

NVIDIA acquires Run:ai, integrates it with DGX Cloud AI Platform 

NVIDIA has acquired Run:ai, an Israeli startup that simplifies AI hardware infrastructure management and optimization for developers and operations teams. The acquisition was made for an undisclosed sum, but sources suggest it was around $700 million.

Run:ai’s platform allows AI models to run parallel across various hardware environments, whether on-premises, in public clouds, or at the edge.

Nvidia plans to maintain Run:ai’s products with their existing business model and will support Run:ai’s product development within Nvidia’s DGX Cloud AI platform. This platform offers enterprise users access to computing infrastructure and software for training AI models, including generative AI.

Why does it matter?

NVIDIA is strengthening its offering across the entire AI stack, from hardware to software. and solidifies its status as a comprehensive solution provider for all your AI infra needs. NVIDIA’s vertical integration strategy aims to simplify and optimize AI deployments for customers, asserting its dominance in the evolving AI landscape.

Source

Snowflake enters the generative AI arena with Arctic LLM

Snowflake, the cloud computing company, has released Arctic LLM, a generative AI model for enterprise use. It’s optimized for generating database code and is available under an Apache 2.0 license.

Arctic LLM outperforms other models like DBRX and Llama3 in tasks like coding and SQL generation. Snowflake aims to address enterprise challenges with this model, including building SQL co-pilots and high-quality chatbots. This move aligns with the trend of cloud vendors offering specialized generative AI solutions for businesses

Why does it matter?

Approximately 46% of global enterprise AI decision-makers use existing open-source LLMs for generative AI. With the release of Arctic, Snowflake democratizes access to cutting-edge models by offering an Apache 2.0 license for ungated personal, research, and commercial use.

Source

Monetizing generative AI to take time, says Zuckerberg

Meta CEO Mark Zuckerberg stated that it would take several years for Meta to make money from generative AI. The company is already profitable, but building advanced AI capabilities will be lengthy and costly. Monetization strategies include scaling business messaging, introducing ads or paid content, and offering larger AI models for a fee. However, it will take time for these efforts to yield significant profits.

Why does it matter?

Mark Zuckerberg’s statement highlights the challenges and time required to monetize generative AI technologies effectively. It underscores the complexity of developing advanced AI capabilities and the need for substantial investments. Furthermore, it emphasizes the importance of long-term planning and patient investment in developing and commercializing AI applications.

Source

🤖 AI start-up unveils avatars that convincingly show human emotions

  • An AI startup named Synthesia has created hyperrealistic AI-generated avatars that are extremely lifelike and expressive, pushing the boundaries of generative AI technology.
  • The avatars can replicate human emotions and mannerisms closely, thanks to advancements in AI and extensive data from human actors, aiming to make digital clones indistinguishable from real humans in videos.
  • Despite the technological marvel, the creation of such realistic avatars raises significant ethical concerns about distinguishing between real and AI-generated content, potentially affecting trust and truth in digital media.
  • Source

🔍 Microsoft and Amazon’s AI ambitions spark regulatory rumble

  • UK regulators are investigating Microsoft and Amazon’s investments in AI startups, such as Amazon’s partnership with Anthropic and Microsoft’s dealings with Mistral AI and Inflection AI, for potential anti-competitive impacts.
  • The CMA is analyzing if these partnerships align with UK merger rules and their effect on competition, following significant investments and strategic hiring by the companies.
  • Both Microsoft and Amazon assert that their AI investments and partnerships promote competition and are confident in a favorable resolution by regulators.
  • Source

What Else Is Happening in AI on April 25th 2024❗

🎥 Adobe unveils VideoGigagan: AI project upscaling blurry videos to HD

Adobe’s VideoGigagan AI project enhances low-quality videos by upscaling them to higher resolutions, even when the original footage is blurry. It uses automatic adjustments for brightness, contrast, saturation, and sharpness, benefiting brand perception, engagement, and customer satisfaction. (Link)

📱 OpenELM: Apple’s evolving AI strategy for iPhones

Apple has unveiled OpenELM, a collection of compact language models that enable AI functionality on its devices. These models, available in four sizes ranging from 270 million to 3 billion parameters, are specifically designed to excel in text-related tasks like email composition.

Just as Google, Samsung and Microsoft continue to push their efforts with generative AI on PCs and mobile devices, Apple is moving to join the party with OpenELM, a new family of open source large language models (LLMs) that can run entirely on a single device rather than having to connect to cloud servers.

There are eight OpenELM models in total – four pre-trained and four instruction-tuned – covering different parameter sizes between 270 million and 3 billion parameters (referring to the connections between artificial neurons in an LLM, and more parameters typically denote greater performance and more capabilities, though not always).

Apple is offering the weights of its OpenELM models under what it deems a “sample code license,” along with different checkpoints from training, stats on how the models perform as well as instructions for pre-training, evaluation, instruction tuning and parameter-efficient fine tuning.

(Link)

☁️ IBM acquires HashiCorp for $6.4 Billion to boost cloud business

IBM has acquired HashiCorp, Inc., for $6.4 billion, aiming to enhance its hybrid cloud and AI capabilities. The acquisition will integrate HashiCorp’s suite of products, including Terraform, to automate hybrid and multi-cloud environments. (Link)

🙂 Synthesia Introduces Emotions to AI Video Avatars

Synthesia, an AI startup specializing in video avatars for business users, has released an update introducing emotions to its avatars. The latest version includes avatars built from actual humans, providing better lip tracking, more expressive natural movements, and improved emotional range when generating videos. (Link)

🤖 HubSpot introduces cutting-edge AI tools for SMBs

HubSpot introduced HubSpot AI at INBOUND 2023, featuring AI assistants for email drafting and content creation, AI agents for customer service, predictive analytics, and ChatSpot powered by OpenAI’s ChatGPT. The revamped Sales Hub offers modernized sales processes tailored for SMBs. (Link)

A Daily chronicle of AI Innovations April 24th 2024: 🖼️ Firefly 3: Adobe’s best AI image generation model to date 👓 Meta finally rolls out multimodal AI capabilities for its smart glasses 🧬 Profulent’s OpenCRISPR-1 can edit the human genome 🤝 Coca-Cola and Microsoft partner to accelerate cloud and Gen AI initiatives 👥 Cognizant and Microsoft team up to boost Gen AI adoption 🧠 Amazon wishes to host companies’ custom Gen AI models 🚀 OpenAI launches more enterprise-grade features for API customers 🤖 Tesla could start selling Optimus robots by the end of 2025 ❄️ Snowflake launches 480bn-parameter AI to take on OpenAI, Google and Meta 👓 Meta adds AI to its Ray-Ban smart glasses 📉 Apple reduces production of Vision Pro due to low demand

Firefly 3: Adobe’s best AI image generation model to date 

Adobe has announced a major update to its AI image generation technology called Firefly Image 3. The model showcases a significant improvement in creating more realistic and high-quality images over previous versions. It has enhanced capabilities to understand longer text prompts, generate better lighting, and depict subjects like crowds and human expressions. The Firefly Image 3 model is now available through Adobe’s Firefly web app as well as integrated into Adobe Photoshop and InDesign apps.

It powers new AI-assisted features in these apps, such as generating custom backgrounds, creating image variations, and enhancing detail. Adobe has also introduced advanced creative controls like Structure Reference to match a reference image’s composition and Style Reference to transfer artistic styles between images. Adobe also attaches “Content Credentials” to all Firefly-generated assets to promote responsible AI development.

Why does it matter?

In AI image generation, a more powerful model from a major player like Adobe could intensify competition with rivals like Midjourney and DALL-E It may motivate other providers to accelerate their own model improvements to keep pace. For creative professionals and enthusiasts, accessing such advanced AI tools could unlock new levels of creative expression and productivity.

Source

Meta finally rolls out multimodal AI capabilities for its smart glasses; adds new features

Meta has announced exciting updates to their Ray-Ban Meta smart glasses collection. They are introducing new styles to cater to a wider range of face shapes. The new styles include the vintage-inspired Skyler frames, designed for smaller faces, and the Headliner frames with a low bridge option. It also introduces video calling capabilities via WhatsApp and Messenger, allowing users to share their views during a video call.

Meta is integrating its AI technology, Meta AI Vision, into Ray-Ban smart glasses. Users can interact with the glasses using voice commands, saying “Hey Meta,” and receive real-time information. The multimodal AI can translate text into different languages using the built-in camera.  These capabilities were in testing for a while and are now available to everyone in the US and Canada.

Why does it matter?

Meta is pushing the boundaries of smart glasses technology, making them more versatile, user-friendly, and AI-powered. This could lead to increased mainstream adoption and integration of augmented reality wearables and voice-controlled AI assistants. Smart glasses could also redefine how people interact with the world around them, potentially changing how we work, communicate, and access information in the future.

Source

Profulent’s OpenCRISPR-1 can edit the human genome

Profluent, a biotechnology company, has developed the world’s first precision gene editing system using AI-generated components. They trained LLMs on a vast dataset of CRISPR-Cas proteins to generate novel gene editors that greatly expand the natural diversity of these systems. OpenCRISPR-1 performed similarly to the widely used SpCas9 gene editor regarding on-target editing activity but had a 95% reduction in off-target effects. This means OpenCRISPR-1 can edit the human genome with high precision.

Profulent’s OpenCRISPR-1 can edit the human genome
Profulent’s OpenCRISPR-1 can edit the human genome

The researchers further improved OpenCRISPR-1 by using AI to design compatible guide RNAs, enhancing its editing efficiency. Profluent publicly released OpenCRISPR-1 to enable broader, ethical use of this advanced gene editing technology across research, agriculture, and therapeutic applications. By using AI-generated components, they aim to lower the cost and barriers to accessing powerful genome editing capabilities.

Why does it matter?

The ability to design custom gene editors using AI could dramatically accelerate the pace of innovation in gene editing, making these powerful technologies more precise, safer, accessible, and affordable for a wide range of diseases. This could lead to breakthroughs like personalized medicine, agricultural applications, and basic scientific research.

Source

What Else Is Happening in AI on April 24th 2024❗

🤝 Coca-Cola and Microsoft partner to accelerate cloud and Gen AI initiatives

Microsoft and Coca-Cola announced a 5-year strategic partnership, where Coca-Cola has made a $1.1 billion commitment to the Microsoft Cloud and its generative AI capabilities. The collaboration underscores Coca-Cola’s ongoing technology transformation, underpinned by the Microsoft Cloud as Coca-Cola’s globally preferred and strategic cloud and AI platform. (Link)

👥 Cognizant and Microsoft team up to boost Gen AI adoption 

Microsoft has teamed up with Cognizant to bring Microsoft’s Gen AI capabilities to Cognizant’s employees and users. Cognizant acquired 25,000 Microsoft 365 Copilot seats for its associates, 500 Sales Copilot seats, and 500 Services Copilot seats. With that, Cognizant will transform business operations, enhance employee experiences, and deliver new customer value. (Link)

🧠 Amazon wishes to host companies’ custom Gen AI models

AWS wants to become the go-to place for companies to host and fine-tune their custom Gen AI models. Amazon Bedrock’s new Custom Model Import feature lets organizations import and access Gen AI models as fully managed APIs. Companies’ proprietary models, once imported, benefit from the same infrastructure as other generative AI models in Bedrock’s library. (Link)

🚀 OpenAI launches more enterprise-grade features for API customers 

OpenAI expanded its enterprise features for API customers, further enriching its Assistants API and introducing new tools to enhance security and administrative control. The company has introduced Private Link, a secure method to enable direct communication between Azure and OpenAI. It has also added Multi-Factor Authentication (MFA) to bolster access control. (Link)

🤖 Tesla could start selling Optimus robots by the end of 2025

According to CEO Elon Musk, Tesla’s humanoid robot, Optimus, may be ready to sell by the end of next year. Several companies have been betting on humanoid robots to meet potential labor shortages and perform repetitive tasks that could be dangerous or tedious in industries such as logistics, warehousing, retail, and manufacturing. (Link))

📱Microsoft launches Phi-3, its smallest AI model yet

Microsoft launched the next version of its lightweight AI model Phi-3 Mini, the first of three small models the company plans to release.

The company released Phi-2 in December, which performed just as well as bigger models like Llama 2.

Eric Boyd, corporate vice president of Microsoft Azure AI Platform, tells The Verge Phi-3 Mini is as capable as LLMs like GPT-3.5 “just in a smaller form factor.”

Compared to their larger counterparts, small AI models are often cheaper to run and perform better on personal devices like phones and laptops.

Source

👓 Meta adds AI to its Ray-Ban smart glasses

  • Ray-Ban Meta smart glasses now include multimodal AI, enabling the device to process diverse types of data such as images, videos, text, and sound to understand the user’s environment in real-time.
  • The AI capabilities allow users to interact with their surroundings in enhanced ways, such as identifying dog breeds, translating signs in foreign languages, and offering recipe suggestions based on visible ingredients.
  • Initial testing of the multimodal AI has shown promise, although it has also revealed some inconsistencies in accuracy, such as errors in identifying certain car models and plant species.
  • Source

📉 Apple reduces production of Vision Pro due to low demand

  • Apple is reducing production of its Vision Pro headset for the rest of 2024 due to lower than expected demand, with sales projections adjusted down from up to 800,000 units to around 400,000 to 450,000 units.
  • Following weaker sales and reduced demand, the launch of a more affordable mixed-reality headset from Apple could be delayed until after 2025, as the company reassesses its Vision Pro strategy.
  • Despite efforts to boost Vision Pro’s appeal, including introducing new features and accessories, lack of key app support and customer dissatisfaction with practicality are contributing to its sluggish sales.
  • Source

❄️ Snowflake launches 480bn-parameter AI to take on OpenAI, Google and Meta 

  • Snowflake announced Arctic LLM, an enterprise-grade generative AI model designed for generating database code and available under an Apache 2.0 license for free commercial and research use.
  • Arctic LLM, using a mixture of experts (MoE) architecture, claims to outperform competitors like DBRX and certain models from Meta on coding and SQL generation tasks.
  • Snowflake aims to integrate Arctic LLM into its platform, Cortex, offering it as a solution for building AI- and machine learning-powered apps with a focus on security, governance, and scalability.
  • Source

Discover the Ultimate AI Tools List

Natural Language Processing (NLP):

  1. OpenAI GPT (Generative Pre-trained Transformer)
  2. Google Cloud Natural Language API
  3. SpaCy
  4. MyEssayWriter.ai
  5. NLTK (Natural Language Toolkit)
  6. AllenNLP

Computer Vision:

  1. TensorFlow
  2. OpenCV (Open Source Computer Vision Library)
  3. PyTorch
  4. YOLO (You Only Look Once)
  5. Caffe

Speech Recognition:

  1. Google Cloud Speech-to-Text
  2. IBM Watson Speech to Text
  3. CMU Sphinx (PocketSphinx)
  4. Kaldi
  5. Mozilla DeepSpeech

Machine Learning Frameworks:

  1. TensorFlow
  2. PyTorch
  3. scikit-learn
  4. Keras
  5. Microsoft Azure Machine Learning

Chatbots and Conversational AI:

  1. Dialogflow
  2. IBM Watson Assistant
  3. Microsoft Bot Framework
  4. Rasa
  5. Amazon Lex

Data Analytics and Visualization:

  1. Tableau
  2. Power BI
  3. Google Data Studio
  4. Plotly
  5. Matplotlib

AI Development Platforms:

  1. H2O.ai
  2. DataRobot
  3. RapidMiner
  4. Domino Data Lab
  5. Dataiku

Reinforcement Learning:

  1. OpenAI Gym
  2. Stable Baselines
  3. RLlib (Reinforcement Learning Library)

AI Ethics and Bias Mitigation:

  1. IBM AI Fairness 360
  2. Google’s What-If Tool
  3. Microsoft Fairlearn

Generative Adversarial Networks (GANs):

  1. NVIDIA StyleGAN
  2. CycleGAN
  3. Pix2Pix

Automated Machine Learning (AutoML):

  1. Auto-Keras
  2. Google Cloud AutoML
  3. H2O.ai Driverless AI
  4. TPOT (Tree-based Pipeline Optimization Tool)
  5. Auto-Sklearn

Time Series Forecasting:

  1. Statsmodels
  2. ARIMA (AutoRegressive Integrated Moving Average)
  3. LSTM (Long Short-Term Memory) networks
  4. XGBoost

Optimization and Operations Research:

  1. IBM CPLEX
  2. Gurobi
  3. Pyomo
  4. Google OR-Tools

Knowledge Graphs:

  1. Neo4j
  2. Amazon Neptune
  3. Stardog
  4. Ontotext GraphDB

AI Infrastructure and Deployment:

  1. Kubernetes
  2. Docker
  3. AWS SageMaker
  4. Google Cloud AI Platform
  5. Microsoft Azure Machine Learning Service

Text Analysis and Sentiment Analysis:

  • VADER (Valence Aware Dictionary and sEntiment Reasoner)
  • TextBlob
  • IBM Watson Natural Language Understanding
  • Lexalytics
  • Aylien Text Analysis API

Recommendation Systems:

  1. Apache Mahout
  2. LightFM
  3. Surprise
  4. Amazon Personalize
  5. TensorFlow Recommenders

AI-driven Marketing Tools:

  1. Salesforce Einstein
  2. Marketo
  3. HubSpot
  4. Adobe Sensei
  5. Optimizely

AI-powered Content Creation:

  1. Artbreeder
  2. Copy.ai
  3. ShortlyAI
  4. Jasper (Journalism AI)
  5. AI Dungeon
  6. PerfectEssayWriter.ai
  7. MyPerfectPaper.net – AI Essay Writing

Healthcare AI Tools:

  1. IBM Watson Health
  2. NVIDIA Clara
  3. Google Health
  4. Ada Health
  5. PathAI

AI in Finance:

  1. AlphaSense
  2. QuantConnect
  3. Kensho Technologies
  4. FactSet
  5. Yewno|Edge

AI in Cybersecurity:

  1. Darktrace
  2. Cylance
  3. CrowdStrike Falcon
  4. Symantec AI Solutions
  5. FireEye Helix

AI in Robotics:

  1. ROS (Robot Operating System)
  2. NVIDIA Isaac
  3. Universal Robots
  4. SoftBank Robotics
  5. Boston Dynamics

AI in Energy and Sustainability:

  1. Google DeepMind for Energy
  2. C3.ai
  3. GridGain Systems
  4. Siemens Digital Grid
  5. Envision Digital

AI in Agriculture:

  1. Climate Corporation
  2. Blue River Technology
  3. PrecisionHawk
  4. AgShift
  5. Taranis

AI in Education:

  1. Duolingo
  2. Coursera
  3. Gradescope
  4. DreamBox Learning
  5. Carnegie Learning

AI in Supply Chain Management:

  1. Llamasoft
  2. Blue Yonder (formerly JDA Software)
  3. Element AI
  4. ClearMetal
  5. Kinaxis

AI in Gaming:

  1. Unity ML-Agents
  2. NVIDIA Deep Learning Super Sampling (DLSS)
  3. Unreal Engine AI
  4. Microsoft Project Malmo
  5. IBM Watson Unity SDK

AI in Transportation:

  1. Waymo
  2. Tesla Autopilot
  3. Uber ATG (Advanced Technologies Group)
  4. Didi Chuxing AI Labs
  5. Mobileye by Intel

AI in Customer Service:

  1. Zendesk AI
  2. Ada Support
  3. Helpshift
  4. Intercom
  5. Freshworks AI

AI in Legal Services:

  1. ROSS Intelligence
  2. Luminance
  3. Kira Systems
  4. Casetext
  5. Lex Machina

AI in Real Estate:

  1. Zillow
  2. Redfin
  3. CompStak
  4. Skyline AI
  5. Matterport

AI in Human Resources:

  1. HireVue
  2. Textio
  3. Pymetrics
  4. Traitify
  5. Visage

AI in Retail:

  1. Amazon Go
  2. Salesforce Commerce Cloud Einstein
  3. Blue Yonder (formerly JDA Software)
  4. Dynamic Yield
  5. Sentient Ascend

AI in Personalization and Recommendation:

  1. Netflix Recommendation System
  2. Spotify Discover Weekly
  3. Amazon Product Recommendations
  4. YouTube Recommendations
  5. Pandora Music Genome Project

AI in Natural Disaster Prediction:

  1. One Concern
  2. Jupiter
  3. Descartes Labs
  4. Zizmos
  5. Earth AI

AI in Language Translation:

  1. Google Translate
  2. DeepL
  3. Microsoft Translator
  4. SYSTRAN
  5. Translate.com

AI in Facial Recognition:

  1. Amazon Rekognition
  2. Face++ by Megvii
  3. Kairos
  4. Microsoft Azure Face API
  5. NEC NeoFace

AI in Music Generation:

  1. AIVA
  2. Amper Music
  3. Jukedeck
  4. Magenta by Google
  5. OpenAI Jukebox

AI in Remote Sensing:

  1. Orbital Insight
  2. Descartes Labs
  3. SkyWatch
  4. TerrAvion
  5. Planet Labs

AI in Document Management:

  1. DocuSign
  2. Adobe Acrobat
  3. Abbyy FineReader
  4. DocuWare
  5. Nitro

AI in Social Media Analysis:

  1. Brandwatch
  2. Sprinklr
  3. Talkwalker
  4. Hootsuite Insights
  5. Synthesio

AI in Fraud Detection:

  1. Feedzai
  2. Forter
  3. Simility
  4. Featurespace
  5. Signifyd

AI in Smart Cities:

  1. Sidewalk Labs
  2. CityBrain by Alibaba Cloud
  3. Siemens City Performance Tool
  4. StreetLight Data
  5. SmartCone

AI in Mental Health:

  1. Woebot
  2. Wysa
  3. X2AI
  4. Talkspace
  5. Ginger

AI in Music Streaming Services:

  1. Spotify
  2. Apple Music
  3. Pandora
  4. Tidal
  5. Deezer

AI in Journalism:

  1. Automated Insights
  2. Narrativa
  3. Heliograf by The Washington Post
  4. Wordsmith by Automated Insights
  5. RADAR by The Associated Press

AI in Predictive Maintenance:

  1. Uptake
  2. IBM Maximo Asset Performance Management
  3. SAS Predictive Maintenance
  4. Predikto
  5. Augury

AI in 3D Printing:

  1. Autodesk Netfabb
  2. Formlabs PreForm
  3. Stratasys GrabCAD
  4. Materialise Magics
  5. SLM Solutions

AI in Wildlife Conservation:

  1. ConservationFIT
  2. PAWS (Protection Assistant for Wildlife Security)
  3. Instant Wild
  4. TrailGuard AI
  5. Wildlife Insights

AI in Graphic Design:

  1. Adobe Sensei (Adobe Creative Cloud’s AI platform)
  2. Canva’s Magic Resize
  3. Designhill’s AI Logo Maker
  4. Tailor Brands
  5. Piktochart

A Daily chronicle of AI Innovations April 23rd 2024: 📱 Microsoft launches its smallest AI model that can fit on your phone 💥 Meta opens Quest OS to third-party developers to rival Apple 🎨 Adobe claims its new image generation model is its best yet 🎨 Adobe survey says 50% Americans use generative AI everyday 🏎️ Mercedes-Benz becomes first automaker to sell Level 3 autonomous vehicles in the US 🫠 GPT-4 can exploit zero-day security vulnerabilities all by itself, a new study finds 🎥 Creative Artists Agency (CAA) is testing an AI initiative called CAA Vault 📷 Poetry Camera by Kelin Carolyn Zhang and Ryan Mather generates poems from pictures 🤖 Alethea AI launched expressive AI avatars on Coinbase’s blockchain

Microsoft launches its smallest AI model that can fit on your phone

Microsoft launched Phi-3-Mini, a 3.8 billion parameter language model, as the first of three small models in the Phi-3 series. It is trained on a smaller dataset than larger LLMs like GPT-4 and outperforms models like Meta’s Llama 2 7B and GPT-3.5 on benchmarks like MMLU and MT-bench. The Phi-3 series also includes Phi-3-Small (7B parameters) and Phi-3-Medium (14B parameters), which are more capable than Phi-3-Mini.

Microsoft launches its smallest AI model that can fit on your phone
Microsoft launches its smallest AI model that can fit on your phone

What sets Phi-3-Mini apart is its ability to run locally on mobile devices like the iPhone 14, thanks to its optimized size and innovative quantization techniques. Microsoft’s team took inspiration from how children learn, using a “curriculum” approach to train Phi-3 on synthetic “bedtime stories” and simplified texts. While robust for its size, Phi-3-Mini is limited in storing extensive factual knowledge and is primarily focused on English.

Why does this matter?

Microsoft’s innovative training approach could lead to more effective and efficient model development techniques. However, Phi-3-Mini’s limitations in storing factual knowledge and its English-centric focus highlight the challenges in creating truly comprehensive and multilingual AI systems.

Source

Adobe survey says 50%Americans use generative AI everyday

Adobe surveyed 3,000 consumers on February 15-19, 2024, about their usage of generative AI and found over half of Americans have already used generative AI. The majority believe it helps them be more creative. Adobe’s Firefly has generated 6.5 billion images since its inception last March. Americans use generative AI for research, brainstorming, creating content, searching, summarization, coding, and learning new skills.

Moreover, 41% of Americans expect brands to use AI for personalized shopping, price comparisons, and customer support. Adobe’s data also reveals that online traffic to retail and travel sites has surged, with faster customer service and more creative experiences due to generative AI tools.

Why does this matter?

Gen AI’s usage has increased over time. Many surveys last year found that very less percentage of Americans used ChatGPT. As generative AI tools become more accessible, businesses must embrace this technology faster to deliver experiences that resonate with modern consumers.

Source

Microsoft hired former Meta VP of infrastructure

With the recent addition of Google DeepMind co-founder Mustafa Suleyman to lead Microsoft’s consumer AI division, Microsoft has once again poached a former Meta VP of infrastructure. This strategic hire comes amidst rumors of Microsoft and OpenAI’s plans to construct a $100 billion supercomputer, “Stargate,” to power their AI models.

Jason Taylor oversaw infrastructure for AI, data, and privacy in Meta. He will join Microsoft as the corporate vice president and deputy CTO, tasked with building systems to advance the company’s AI ambitions.

Why does this matter?

Microsoft’s aggressive moves in the AI space highlight the fierce competition among tech giants. As AI systems become increasingly resource-intensive, having the right talent will be vital for delivering cutting-edge AI experiences. In addition to strategic hires, Microsoft is rumored to develop a supercomputer project, which could have far-reaching implications for various industries.

Source

What Else Is Happening in AI on April 23rd 2024❗

🎥 Creative Artists Agency (CAA) is testing an AI initiative called CAA Vault

Hollywood’s leading talent agency allows their A-list clients to create digital clones of themselves. The agency is partnering with AI firms to scan their clients’ bodies, faces, and voices. These AI replicas can reshoot scenes, dubbing, or superimpose onto stunt doubles in film and TV production. CAA is also planning to make this technology available to the entire industry. (Link)

📷 Poetry Camera by Kelin Carolyn Zhang and Ryan Mather generates poems from pictures

Powered by GPT-4, this open-source AI camera allows users to choose from various poetic forms from the scenes it captures. It prioritizes privacy by not digitally saving images or poems. The positive response has led the creators to consider making the Poetry Camera commercially available. (Link)

🤖 Alethea AI launched expressive AI avatars on Coinbase’s blockchain

Their proprietary Emote Engine powers high-fidelity facial animations, body movements, and generative AI capabilities. The platform lets users create AI agents quickly and collaborate with the community. Creators can also monetize their AI agents without centralized censorship or revenue sharing. Alethea AI aims to create an avatar arena featuring full-body animation, voice, and lip-syncing. (Link)

🎤 TikTok is working on a new feature that lets users clone their voice

Discovered in the latest Android app version, this new AI text-to-speech feature will allow users to record their voices, which will then be added to the TikTok Voice Library for others. While the feature is still under development, it’s already raising concerns about potential misuse and spreading misinformation. TikTok is expected to provide additional details on privacy and safety measures when the feature is ready for broader release. (Link)

💥 Meta opens Quest OS to third-party developers to rival Apple

  • Meta is licensing its Horizon OS, designed for Quest headsets, to hardware manufacturers such as Lenovo and Asus and creating a special Quest version with Xbox.
  • The company is promoting alternative app stores on its platform, making its App Lab store more visible and inviting Google to integrate the Play Store with Horizon OS.
  • With Horizon OS, Meta aims to create a more open ecosystem similar to Microsoft’s approach with Windows, focusing on expanding its social network Horizon through licensing and hardware partnerships.
  • Source

🎨 Adobe claims its new image generation model is its best yet

  • Adobe has introduced its third-generation image-generation model, Firefly Image 3, which boasts enhanced realism and improved rendering capabilities for complex scenes and lighting, compared to its predecessors.
  • Firefly Image 3, which is now integrated into Photoshop and the Adobe Firefly web app, features advancements such as better understanding of detailed prompts, more accurate depiction of dense crowds, and improved text and iconography rendering.
  • In addition to technical improvements, Adobe emphasizes ethical AI practices with Firefly Image 3 by using a diverse and ethically sourced training dataset, including content from Adobe Stock and AI-generated images under strict moderation.
  • Source

 Mercedes-Benz becomes first automaker to sell Level 3 autonomous vehicles in the US

  • Mercedes-Benz is the first automaker to sell Level 3 autonomous driving vehicles in the U.S., with the EQS and S-Class sedans now available in California and Nevada.
  • The Drive Pilot feature in these vehicles allows drivers to take their eyes off the road and hands off the wheel in certain conditions, requiring a $2,500 yearly subscription.
  • Drive Pilot can be activated only during specific conditions such as clear weather, daytime, in heavy traffic under 40 mph, and on preapproved freeways in California and Nevada.
  • Source

🫠 GPT-4 can exploit zero-day security vulnerabilities all by itself, a new study finds

  • GPT-4 has demonstrated the ability to exploit zero-day security vulnerabilities autonomously, as revealed by a new study.
  • The study, conducted by researchers from the University of Illinois Urbana-Champaign, found that GPT-4 could exploit 87% of tested vulnerabilities, significantly outperforming other models including GPT-3.5.
  • Despite the potential for “security through obscurity” strategies, the researchers advocate for more proactive security measures against the risks posed by highly capable AI agents like GPT-4.
  • Source

A Daily chronicle of AI Innovations April 22 2024: 🍎 iOS 18 to have AI features with on-device processing  🧠 Many-shot ICL is a breakthrough in improving LLM performance  ⚡ Groq shatters AI inference speed record with 800 tokens/second on LLaMA 3 🤖 Why Zuckerberg wants to give away a $10B AI model 🤐 Sundar Pichai tells Google staff he doesn’t want any more political debates in the office 🤖 Israel-based startup enters AI humanoid race with Menteebot 🩺 Hugging Face introduces benchmark for evaluating gen AI in healthcare 🔄 Google announces major restructuring to accelerate AI development 🎧 Nothing’s new earbuds offer ChatGPT integration 🚪 Japanese researchers develop AI tool to predict employee turnover

iOS 18 to have AI features with complete on-device processing

Apple is set to make significant strides in artificial intelligence with the upcoming release of iOS 18. According to Apple Insider’s recent report, the tech giant is focusing on privacy-centric AI features that will function entirely on-device, eliminating the need for cloud-based processing or an internet connection. This approach addresses concerns surrounding AI tools that rely on server-side processing, which have been known to generate inaccurate information and compromise user privacy.

The company is reportedly developing an in-house LLM called “Ajax,” which will power AI features in iOS 18. Users can expect improvements to Messages, Safari, Spotlight Search, and Siri, with basic text analysis and response generation available offline. We’ll learn more about Apple’s AI plans at the Worldwide Developers Conference (WWDC) starting June 10.

Why does this matter?

Apple’s commitment to user data privacy is commendable, but eliminating cloud-based processing and internet connectivity may impede the implementation of more advanced features. Nevertheless, it presents an opportunity for Apple to differentiate itself from competitors by offering users a choice between privacy-focused on-device processing and more powerful cloud-based features.

Source

Many-shot-in-context learning is a breakthrough in improving LLM performance

A recent research paper has introduced a groundbreaking technique that enables LLMs to significantly improve performance by learning from hundreds or thousands of examples provided in context. This approach, called many-shot in-context learning (ICL), has shown superior results compared to the traditional few-shot learning method across a wide range of generative and discriminative tasks.

Many-shot-in-context learning is a breakthrough in improving LLM performance
Many-shot-in-context learning is a breakthrough in improving LLM performance

To address the limitation of relying on human-generated examples for many-shot ICL, the researchers explored two novel settings: Reinforced ICL, which uses model-generated chain-of-thought rationales instead of human examples, and Unsupervised ICL, which removes rationales from the prompt altogether and presents the model with only domain-specific questions.

Both approaches have proven highly effective in the many-shot regime, particularly for complex reasoning tasks. Furthermore, the study reveals that many-shot learning can override pretraining biases and learn high-dimensional functions with numerical inputs, unlike few-shot learning, showcasing its potential to revolutionize AI applications.

Why does this matter?

Many-shot ICL allows for quick adaptation to new tasks and domains without the need for extensive fine-tuning or retraining. However, the success of many-shot ICL heavily depends on the quality and relevance of the examples provided. Moreover, as shown by Anthropic’s jailbreaking experiment, some users could use this technique to intentionally provide carefully crafted examples designed to exploit vulnerabilities or introduce biases, leading to unintended and dangerous consequences.

Source

Groq shatters AI inference speed record with 800 tokens/second on LLaMA 3

AI chip startup Groq recently confirmed that its novel processor architecture is serving Meta’s newly released LLaMA 3 large language model at over 800 tokens per second. This translates to generating about 500 words of text per second – nearly an order of magnitude faster than the typical speeds of large models on mainstream GPUs. Early testing by users seems to validate the claim.

Groq’s Tensor Streaming Processor is designed from the ground up to accelerate AI inference workloads, eschewing the caches and complex control logic of general-purpose CPUs and GPUs. The company asserts this “clean sheet” approach dramatically reduces the latency, power consumption, and cost of running massive neural networks.

Why does this matter?

If the LLaMA 3 result holds up, it could shake up the competitive landscape for AI inference, challenging Nvidia’s dominance of GPUs and increasing the demand for purpose-built AI hardware for faster and more cost-effective inference solutions. Also, Groq’s capabilities could revolutionize software solutions that depend on real-time AI, such as virtual assistants, chatbots, and interactive customer services.

Source

🤖 Why Zuckerberg wants to give away a $10B AI model 

  • Mark Zuckerberg, CEO of Meta, said in a podcast he would be willing to open source a $10 billion AI model under certain conditions if it was safe and beneficial for all involved.
  • Zuckerberg believes that open sourcing can mitigate dependency on a few companies controlling AI technology, fostering innovation and competition.
  • He also points to Meta’s strong open-source legacy with projects like PyTorch and the Open Compute Project, which have significantly reduced costs and expanded supply chains by making their designs available to the public.
  • Source

🤐 Sundar Pichai tells Google staff he doesn’t want any more political debates in the office

  • Google CEO Sundar Pichai directed employees to stop bringing political debates into the workplace, emphasizing the company as a business focused on being an objective information provider.
  • The directive came after 28 employees were fired for protesting against a controversial cloud computing contract.
  • Pichai’s stance reflects a broader trend in tech companies to restrict political discourse at work to maintain focus and avoid internal conflicts, with companies like Coinbase and Meta implementing similar policies.
  • Source

What Else Is Happening in AI on April 22 2024❗

🤖 Israel-based startup enters AI humanoid race with Menteebot

Israel-based startup Mentee Robotics has unveiled Menteebot, an AI-driven humanoid robot prototype for home and warehouse use. It employs transformer-based large language models, NeRF-based algorithms, and simulator-to-reality machine learning to understand commands, create 3D maps, and perform tasks. The finalized Menteebot is anticipated to launch in Q1 2025. (Link)

🩺 Hugging Face introduces benchmark for evaluating gen AI in healthcare

The benchmark combines existing test sets to assess medical knowledge and reasoning across various fields. It’s a starting point for evaluating healthcare-focused AI models, but experts caution against relying solely on the benchmark and emphasize the need for thorough real-world testing. (Link)

🔄 Google announces major restructuring to accelerate AI development

The changes involve consolidating AI model building at Google Research and DeepMind, focusing Google Research on foundational breakthroughs and responsible AI practices, and introducing a new “Platforms & Devices” product area. (Link)

🎧 Nothing’s new earbuds offer ChatGPT integration

Nothing Ear and Nothing Ear (a) allow users to ask questions by pinching the headphones’ stem, provided the ChatGPT app is installed on a connected Nothing handset. The earbuds offer improved sound quality, better noise-canceling, and longer battery life than their predecessors. (Link)

🚪 Japanese researchers develop AI tool to predict employee turnover

The tool analyzes employee data, such as attendance records and personal information, and creates a turnover model for each company. By predicting which new recruits are likely to quit, the AI tool enables managers to offer targeted support to those employees and potentially reduce turnover rates. (Link)

Apple acquires Paris-based AI company Datakalab to bolster its AI technology. LINK

China’s AI data centers to outpace Korea’s human water consumption. LINK

Google Gemini app on Android may soon let you read ‘real-time responses’. LINK

Bitcoin miners upgrade power centers and get into AI to brace for slashed revenue post halving. LINK

Ecosia launches world’s first energy-generating browser. LINK

A Daily chronicle of AI Innovations April 20th 2024 and April Week 3 Recap: 🤖 OpenAI fires back at Elon Musk 🧠 Google DeepMind researchers call for limits on AI that mimics humans 💰 Bitcoin just completed its fourth-ever ‘halving’ 🚫 Twitter alternative Post News is shutting down

🤖 OpenAI fires back at Elon Musk 

  • OpenAI has refuted Elon Musk’s lawsuit allegations, asserting that he is attempting to discredit the company for his own commercial gain after a failed attempt to dominate it years ago.
  • The company’s legal team disputes Musk’s claim that OpenAI violated its founding principles by commercializing its technology and forming a partnership with Microsoft.
  • OpenAI has requested a court to dismiss Musk’s lawsuit, arguing that there is no basis for his claims, and a hearing for the motion is set for April 24.
  • Source

🧠 Google DeepMind researchers call for limits on AI that mimics humans 

  • Google DeepMind researchers advocate for setting limits on AI that imitates human behaviors, highlighting the risk of users forming overly close bonds that could lead to loss of autonomy and disorientation.
  • Their paper discusses the potential of AI assistants to enhance daily life by acting as partners in creativity, analysis, and planning, but warns they could also misalign with user and societal interests, potentially exacerbating social and technological inequalities.
  • Researchers call for comprehensive research and protective measures, including restrictions on human-like elements in AI, to ensure these systems preserve user autonomy and prevent negative social impacts while promoting the advancement of socially beneficial AI.
  • Source

What happened in AI from April 14th to April 20th 2024❗

📊 xAI’s first multimodal model with a unique dataset
♾️ Infini-Attention: Google’s breakthrough gives LLMs limitless context
⚠️ Adobe’s Firefly AI trained on competitor’s images: Bloomberg report
🎬 Adobe partners with OpenAI, RunwayML & Pika for Premiere Pro
🚀 Reka launches Reka Core: Their frontier in multimodal AI
🏢 OpenAI is opening its first international office in Tokyo
🎮 NVIDIA RTX A400 A1000: Lower-cost single slot GPUs
🎵 Amazon Music launches Maestro, an AI-based playlist generator
💼 Stanford’s report reflects industry dominance and rising training costs in AI
👤 Microsoft VASA-1 generates lifelike talking faces with audio
🤖 Boston Dynamics charges up for the future by electrifying Atlas
🧠 Intel reveals world’s largest brain-inspired computer
🦙 Meta released two Llama 3 models; 400B+ models in training
📈 Mixtral 8x22B claims highest open-source performance and efficiency
🦈 Meta’s Megalodon to solve the fundamental challenges of the Transformer

A Daily chronicle of AI Innovations April 19th 2024: ⚔️ Meta declares war on OpenAI 🦙 Meta’s Llama 3 models are here; 400B+ models in training! 🤖 Google consolidates teams with aim to create AI products faster 🚫 Apple pulls WhatsApp, Threads and Signal from app store in China 🦠 Moderna CEO says AI will help scientists understand ‘most diseases’ in 3 to 5 years 📈 Mixtral 8x22B claims highest open-source performance and efficiency 🦈 Meta’s Megalodon to solve the fundamental challenges of the Transformer 🔍Meta adds its AI chatbot, powered by Llama 3, to the search bar in all its apps. 🚗Wayve introduces LINGO-2, a groundbreaking AI model that drives and narrates its journey. 🤖Salesforce updates Slack AI with smart recaps and more languages. ✈️US Air Force tests AI-controlled jets against human pilots in simulated dogfights. 🔋Google Maps will use AI to find out-of-the-way EV chargers for you.

Meta’s Llama 3 models are here; 400B+ models in training!

Llama 3 is finally here! Meta introduced the first two models of the Llama 3 family for broad use: pretrained and instruction-fine-tuned language models with 8B and 70B parameters. Meta claims these are the best models existing today at the 8B and 70B parameter scale, with greatly improved reasoning, code generation, and instruction following, making Llama 3 more steerable.

Meta’s Llama 3 models are here; 400B+ models in training!
Meta’s Llama 3 models are here; 400B+ models in training!

But that’s not all. Meta is also training large models with over 400B parameters. Over coming months, it will release multiple models with new capabilities including multimodality, the ability to converse in multiple languages, a much longer context window, and stronger overall capabilities.

Meta’s Llama 3 models are here; 400B+ models in training!
Meta’s Llama 3 models are here; 400B+ models in training!

Llama 3 models will soon be available on AWS, Databricks, Google Cloud, Hugging Face, Kaggle, IBM WatsonX, Microsoft Azure, NVIDIA NIM, and Snowflake, and with support from hardware platforms offered by AMD, AWS, Dell, Intel, NVIDIA, and Qualcomm.

Why does this matter?

While Llama 400B+ is still in training, it is already trending. Its release might mark a watershed moment for AI as the open-source community gains access to a GPT4-class model. It will be a powerful foundation for research efforts, and it could be a win for open-source in the longer run if startups/businesses start building more local, tailored models with it.

Source

Mixtral 8x22B claims highest open-source performance and efficiency

Mistral AI has unveiled Mixtral 8x22B, a new open-source language model that the startup claims achieves the highest open-source performance and efficiency. it’s sparse mixture-of-experts (SMoE) model actively uses only 39 billion of its 141 billion parameters. As a result, it offers an exceptionally good price/performance ratio for its size.

The model’s other strengths include multilingualism, with support for English, French, Italian, German, and Spanish, as well as strong math and programming capabilities.

Mixtral 8x22B claims highest open-source performance and efficiency
Mixtral 8x22B claims highest open-source performance and efficiency

Why does this matter?

While Mistral AI’s claims may be true, there’s a new competitor on the market- Llama 3. So we might have to reconsider the claims on the best open-source model out right now. But whatever the benchmarks say, only the practical usefulness of these models will tell which is truly superior.

Source

Meta’s Megalodon to solve the fundamental challenges of the Transformer

Researchers at Meta and the University of Southern California have proposed a new model that aims to solve some of the fundamental challenges of the Transformer, the deep learning architecture that gave rise to the age of LLMs.

The model, called Megalodon, allows language models to extend their context window to millions of tokens without requiring huge amounts of memory. Experiments show that Megalodon outperforms Transformer models of equal size in processing large texts. The researchers have also obtained promising results on small– and medium-scale experiments on other data modalities and will later work on adapting Megalodon to multi-modal settings.

Why does this matter?

Scientists have been looking for alternative architectures that can replace transformers. Megalodon is the latest in the series. However, much research has already been poured into enhancing and making transformers efficient. For example, Google’s Infini-attention released this week. So, the alternatives have a lot to catch up to. For now, transformers continue to remain the dominant architecture for language models.

Source

⚔️ Meta declares war on OpenAI 

  • Meta has expanded the integration of its AI assistant into platforms like Instagram, WhatsApp, Facebook, and a standalone website, aiming to challenge ChatGPT in the AI chatbot market.
  • Meta announced Llama 3, its latest AI model, which reportedly outperforms its predecessors and competitors in several benchmarks, with versions available for both internal use and external developers.
  • CEO Mark Zuckerberg stated that with Llama 3, Meta aims to establish the most advanced and globally accessible AI assistant, featuring enhanced capabilities such as integrated real-time search results and improved image generation.
  • Source

🤖 Google consolidates teams with aim to create AI products faster

  • Google is merging its Android and Chrome software division with the Pixel and Fitbit hardware division to more extensively incorporate artificial intelligence into the company.
  • CEO Sundar Pichai stated that this integration aims to “turbocharge the Android and Chrome ecosystems” and foster innovation under the leadership of executive Rick Osterloh.
  • The reorganization reflects Google’s strategy to leverage AI for consumer and enterprise applications, emphasizing AI’s role in enhancing features like the Pixel camera.
  • Source

🦠 Moderna CEO says AI will help scientists understand ‘most diseases’ in 3 to 5 years

  • Moderna CEO Stéphane Bancel predicted that AI will enable scientists to understand most diseases within the next 3 to 5 years, marking a significant milestone for human health.
  • AI is expected to accelerate drug development, allowing pharmaceutical companies to bring new medicines to patients faster and improving the diagnosis of conditions like heart disease.
  • Bancel expressed optimism about AI’s potential in healthcare, citing insights gained from AI that were previously unknown to scientists, indicating a bright future for medical research and treatment.
  • Source

What Else Is Happening in AI on April 19th 2024❗

🔍Meta adds its AI chatbot, powered by Llama 3, to the search bar in all its apps.

Meta has upgraded its AI chatbot with its newest LLM Llama 3 and has added it to the search bar of its apps– Facebook, Messenger, Instagram, and WhatsApp– in multiple countries. It also launched a new meta.ai site for users to access the chatbot and other new features such as faster image generation and access to web search results. (Link)

🚗Wayve introduces LINGO-2, a groundbreaking AI model that drives and narrates its journey.

LINGO-2 merges vision, language, and action, resulting in every driving maneuver coming with an explanation. This provides a window into the AI’s decision-making, deepening trust and understanding of our assisted and autonomous driving technology. (Link)

🤖Salesforce updates Slack AI with smart recaps and more languages.

Salesforce rolled out Gen AI updates for Slack. The new features build on the native AI smarts– collectively dubbed Slack AI– announced in Feb and provide users with easy-to-digest recaps to stay on top of their day-to-day work interactions. Salesforce also confirmed expanding Slack AI to more languages. (Link)

✈️US Air Force tests AI-controlled jets against human pilots in simulated dogfights.

The Defense Advanced Research Projects Agency (DARPA) revealed that an AI-controlled jet successfully faced a human pilot during an in-air dogfight test last year. The agency has conducted 21 test flights so far and says the tests will continue through 2024. (Link)

🔋Google Maps will use AI to find out-of-the-way EV chargers for you.

Google Maps will use AI to summarize customer reviews of EV chargers to display more specific directions to certain chargers, such as ones in parking garages or more hard-to-find places. The app will also have more prompts to encourage users to submit their feedback after using an EV charger. (Link)

A Daily chronicle of AI Innovations April 18th 2024: 🧠 Samsung unveils lightning-fast DRAM for AI-powered devices 🤖 Logitech’s new AI prompt builder & Signature AI edition mouse 📸 Snapchat to add watermark to images produced with its AI tools ✈️ US Air Force confirms first successful AI dogfight 🏆 Mistral’s latest model sets new records for open source LLMs 🎭 Microsoft’s new AI model creates hyper-realistic video using static image 👁️ GPT-4 nearly matches expert doctors in eye assessments 🔒 Brave unleashes real-time privacy-focused AI answer engine 📸 Snapchat to add watermark to images produced with its AI tools

Microsoft’s VASA-1 generates lifelike talking faces with audio

Microsoft Research’s groundbreaking project, VASA-1, introduces a remarkable framework for generating lifelike talking faces from a single static image and a speech audio clip.

This premiere model achieves exquisite lip synchronization and captures a rich spectrum of facial nuances and natural head motions, resulting in hyper-realistic videos.

Why does it matter?

VASA-1 is crucial in AI for improving lifelike interactions with realistic facial expressions, benefiting customer service, education, and companionship. Its expressive features also enhance storytelling in games and media. Additionally, VASA-1 contributes to developing accessibility tools for those with communication challenges.

Source

Boston Dynamics charges up for the future by electrifying Atlas

Boston Dynamics has unveiled an electric version of their humanoid robot, Atlas. Previously powered by hydraulics, the new Atlas operates entirely on electricity. This development aims to enhance its strength and range of motion, making it more versatile for real-world applications.

Boston Dynamics also plans to collaborate with partners like Hyundai to test and iterate Atlas applications in various environments, including labs, factories, and everyday life.

Why does it matter?

The electric version of Boston Dynamics’ humanoid robot, Atlas, matters because it offers enhanced strength, agility, and practicality for real-world applications. Its electric power source allows it to move in ways that exceed human capabilities, making it versatile for various tasks

Source

Intel reveals world’s largest brain-inspired computer

Intel has introduced the world’s largest neuromorphic computer, mimicking the human brain. Unlike traditional computers, it combines computation and memory using artificial neurons. With 1.15 billion neurons, it consumes 100 times less energy than conventional machines. It performs 380 trillion synaptic operations per second This breakthrough could revolutionize AI and enhance energy-efficient computing.

Why does it matter?

In current AI models, data transfer between processing units can be bottlenecks. Neuromorphic architectures, directly address this issue by integrating computation and memory. This could lead to breakthroughs in training deep learning models.

Source

✈️ US Air Force confirms first successful AI dogfight

  • The US Air Force, via DARPA, announced that an AI-controlled jet successfully engaged in an in-air dogfight against a human pilot for the first time, during tests at Edwards Air Force Base in California in September 2023.
  • DARPA has been working on AI for air combat through its Air Combat Evolution (ACE) program since December 2022, aiming to develop AI capable of autonomously flying fighter jets while adhering to safety protocols.
  • The AI was tested in a real aircraft, the experimental X-62A, against an F-16 flown by a human, achieving close maneuvers without the need for human pilots to intervene, and plans to continue testing through 2024.
  • Source

🏆 Mistral’s latest model sets new records for open source LLMs 

  • French AI startup Mistral AI has released Mixtral 8x22B, claiming it to be the highest-performing and most efficient open-source language model, utilizing a sparse mixture-of-experts model with 39 billion of its 141 billion parameters active.
  • Mixtral 8x22B excels in multilingual support and possesses strong math and programming capabilities, despite having a smaller context window compared to leading commercial models like GPT-4 or Claude 3.
  • The model, licensed under the Apache 2.0 license for unrestricted use, achieves top results on various comprehension and logic benchmarks and outperforms other models in its supported languages on specific tests.
  • Source

🧠 Intel unveils the world’s largest neuromorphic computer

  • Intel Labs introduced its largest neuromorphic computer yet, the Hala Point, featuring 1.15 billion neurons, likened to the brain capacity of an owl, that aims to process information more efficiently by emulating the brain’s neurons and synapses in silicon.
  • The Hala Point system, consuming 2,600 W, is designed to achieve deep neural network efficiencies up to 15 TOPS/W at 8-bit precision, significantly surpassing Nvidia’s current and forthcoming systems in energy efficiency.
  • While showcasing remarkable potential for AI inference and optimization problems with significantly reduced power consumption, Intel’s neuromorphic technology is not yet a universal solution for all AI workloads, with limitations in general-purpose AI acceleration and challenges in adapting large language models.
  • Source

🎭 Microsoft’s new AI model creates hyper-realistic video using static image

  • Microsoft introduced VASA-1, an AI model that produces hyper-realistic videos from a single photo and audio clip, featuring realistic lip syncs and facial movements.
  • The model can create 512x512p resolution videos at 40fps from one image, support modifications like eye gaze and emotional expressions, and even incorporate singing or non-English audio.
  • While Microsoft recognizes the AI’s potential for misuse in creating deepfakes, it intends to use VASA-1 solely for developing virtual interactive characters and advancing forgery detection.
  • Source

👁️ GPT-4 nearly matches expert doctors in eye assessments

  • OpenAI’s GPT-4 almost matched the performance of expert ophthalmologists in an eye assessment study, as reported by the Financial Times and conducted by the University of Cambridge’s School of Clinical Medicine.
  • GPT-4 scored higher than trainee and junior doctors with 60 correct answers out of 87, closely following the expert doctors’ average score of 66.4, in a test evaluating knowledge on various ophthalmology topics.
  • The study, highlighting both potential benefits and risks, indicates that while GPT-4 shows promise in medical assessments, concerns about inaccuracies and the model’s tendency to “hallucinate” answers remain.
  • Source

What Else Is Happening in AI on April 18th 2024❗

🧠 Samsung unveils lightning-fast DRAM for AI-powered devices

Samsung Electronics has achieved a significant milestone by developing the industry’s fastest LPDDR5X DRAM, capable of reaching speeds up to 10.7 Gbps. This new LPDDR5X offers 25% higher performance and 30% more capacity, making it an optimal solution for the on-device AI era. (Link)

🤖 Logitech’s new AI prompt builder & Signature AI edition mouse

Logitech has launched the Logi AI Prompt Builder, a free software tool that enhances interaction with OpenAI’s ChatGPT. It allows Logitech keyboards and mice to serve as shortcuts for more fluent AI prompts. Additionally, Logitech introduced the Signature AI Edition Mouse, featuring a dedicated AI prompt button. (Link)

📸 Snapchat to add watermark to images produced with its AI tools

Snapchat plans to add watermarks to AI-generated images on its platform. These watermarks, featuring a translucent Snap logo and a sparkle emoji, will enhance transparency and prevent content misuse.  (Link)

🔒 Brave unleashes real-time privacy-focused AI answer engine

Brave, the privacy-centric browser, has introduced an AI-driven answer engine within Brave Search. Unlike competitors, it prioritizes privacy by avoiding external search engines. The feature provides real-time generative answers across multiple languages, making it a robust alternative to traditional search.  (Link)

💼 LinkedIn tests premium company page subscription 

LinkedIn is quietly testing a Premium Company Page subscription service for small and medium businesses. The service includes AI-generated content, follower-enhancing tools, and other features to elevate company profiles. Pricing starts at $99.99 per month.    (Link)

A Daily chronicle of AI Innovations April 17th 2024: 🎮 NVIDIA RTX A400 A1000: Lower-cost single slot GPUs; 📊 Stanford’s report reflects industry dominance and rising training costs in AI; 🎵 Amazon Music launches Maestro, an AI playlist generator; 📷 Snap adds watermarks to AI-generated images; 🤖 Boston Dynamics unveils a new humanoid robot; 💰 Andreessen Horowitz raises $7.2 billion, a sign that tech startup market may be bouncing back; 💰 OpenAI offers a 50% discount for off-peak GPT usage; 💻 AMD unveils AI chips for business laptops and desktops; 🧠 Anthropic Claude 3 Opus is now available on Amazon Bedrock; 👤 Zendesk launches an AI-powered customer experience platform; 💼 Intel and The Linux Foundation launch Open Platform for Enterprise AI (OPEA)

Google will pump more than $100B into AI says DeepMind boss

  • DeepMind CEO predicts Google will invest over $100 billion in AI, surpassing rivals like Microsoft in processing prowess.
  • Google’s investment in AI may involve hardware like Axion CPUs based on the Arm architecture, claimed to be faster and more efficient than competitors.
  • Some of the budget will likely go to DeepMind, known for its work on the software side of AI, despite recent mixed results in material discoveries and weather prediction.
  • DeepMind has made progress in teaching AI social skills, a crucial step in advancing AI capabilities.
  • Hassabis emphasized the need for significant computing power, a reason for teaming up with Google in 2014.

Source

A monster of a paper by Stanford, a 500-page report on the 2024 state of AI

Top 10 Takeaways:

  1. AI beats humans on some tasks, but not on all. AI has surpassed human performance on several benchmarks, including some in image classification, visual reasoning, and English understanding. Yet it trails behind on more complex tasks like competition-level mathematics, visual commonsense reasoning and planning.
  2. Industry continues to dominate frontier AI research. In 2023, industry produced 51 notable machine learning models, while academia contributed only 15. There were also 21 notable models resulting from industry-academia collaborations in 2023, a new high.
  3. Frontier models get way more expensive. According to AI Index estimates, the training costs of state-of-the-art AI models have reached unprecedented levels. For example, OpenAI’s GPT-4 used an estimated $78 million worth of compute to train, while Google’s Gemini Ultra cost $191 million for compute.
  4. The United States leads China, the EU, and the U.K. as the leading source of top AI models. In 2023, 61 notable AI models originated from U.S.-based institutions, far outpacing the European Union’s 21 and China’s 15.
  5. Robust and standardized evaluations for LLM responsibility are seriously lacking. New research from the AI Index reveals a significant lack of standardization in responsible AI reporting. Leading developers, including OpenAI, Google, and Anthropic, primarily test their models against different responsible AI benchmarks. This practice complicates efforts to systematically compare the risks and limitations of top AI models.
  6. Generative AI investment skyrockets. Despite a decline in overall AI private investment last year, funding for generative AI surged, nearly octupling from 2022 to reach $25.2 billion. Major players in the generative AI space, including OpenAI, Anthropic, Hugging Face, and Inflection, reported substantial fundraising rounds.
  7. The data is in: AI makes workers more productive and leads to higher quality work. In 2023, several studies assessed AI’s impact on labor, suggesting that AI enables workers to complete tasks more quickly and to improve the quality of their output. These studies also demonstrated AI’s potential to bridge the skill gap between low- and high-skilled workers. Still, other studies caution that using AI without proper oversight can lead to diminished performance.
  8. Scientific progress accelerates even further, thanks to AI. In 2022, AI began to advance scientific discovery. 2023, however, saw the launch of even more significant science-related AI applications— from AlphaDev, which makes algorithmic sorting more efficient, to GNoME, which facilitates the process of materials discovery.
  9. The number of AI regulations in the United States sharply increases. The number of AI related regulations in the U.S. has risen significantly in the past year and over the last five years. In 2023, there were 25 AI-related regulations, up from just one in 2016. Last year alone, the total number of AI-related regulations grew by 56.3%.
  10. People across the globe are more cognizant of AI’s potential impact—and more nervous. A survey from Ipsos shows that, over the last year, the proportion of those who think AI will dramatically affect their lives in the next three to five years has increased from 60% to 66%. Moreover, 52% express nervousness toward AI products and services, marking a 13 percentage point rise from 2022. In America, Pew data suggests that 52% of Americans report feeling more concerned than excited about AI, rising from 37% in 2022.

Source

NVIDIA RTX A400 A1000: Lower-cost single slot GPUs

NVIDIA is expanding its lineup of professional RTX graphics cards with two new desktop GPUs – the RTX A400 and RTX A1000. These new GPUs are designed to bring enhanced AI and ray-tracing capabilities to workstation-class computers. The RTX A1000 GPU is already available from resellers, while the RTX A400 GPU is expected to launch in May.

NVIDIA RTX A400

NVIDIA RTX A400 A1000: Lower-cost single slot GPUs
NVIDIA RTX A400 A1000: Lower-cost single slot GPUs

With 24 tensor cores for AI processing, the A400 enables professionals to run AI apps directly on their desktops, such as intelligent chatbots and copilots. The GPU allows creatives to produce vivid, physically accurate 3D renderings. The A400 also features four display outputs, making it ideal for high-density display environments such as financial services, command and control, retail, and transportation.

NVIDIA RTX A1000

With 72 Tensor Cores, the A1000 offers 3x faster generative AI processing for tools like Stable Diffusion. The A1000 also excels in video processing, as it can process up to 38% more encoding streams and offers up to 2x faster decoding performance than the previous generation. With their slim single-slot design and power consumption of just 50W, the A400 and A1000 GPUs offer impressive features for compact, energy-efficient workstations.

Why does it matter?

NVIDIA RTX A400 and A1000 GPUs provide professionals with cutting-edge AI, graphics, and computing capabilities to increase productivity and unlock creative possibilities. These GPUs can be used by industrial designers, creatives, architects, engineers, healthcare teams, and financial professionals to improve their workflows and achieve faster and more accurate results. With their advanced features and energy efficiency, these GPUs have the potential to impact the future of AI in various industries.

Source

Amazon Music launches Maestro, an AI-based playlist generator

Amazon Music is launching its AI-powered playlist generator, Maestro, following a similar feature introduced by Spotify. Maestro allows users in the U.S. to create playlists by speaking or writing prompts. The AI will then generate a song playlist that matches the user’s input. This feature is currently in beta and is being rolled out to a subset of Amazon Music’s free, Prime, and Unlimited subscribers on iOS and Android.

Like Spotify’s AI playlist generator, Amazon has built safeguards to block inappropriate prompts. However, the technology is still new, and Amazon warns that Maestro “won’t always get it right the first time.”

Why does it matter?

Introducing AI-powered playlist generators could profoundly impact how we discover and consume music in the future. These AI tools can revolutionize music curation and personalization by allowing users to create highly tailored playlists simply through prompts. This trend could increase user engagement, drive more paid subscriptions, and spur further innovation in AI-powered music experiences as companies offer more cutting-edge features.

Source

Standford’s report reflects industry dominance and rising training costs in AI

The AI Index, an independent report by the Stanford Institute for Human-Centered Artificial Intelligence (HAI), provides a comprehensive overview of global AI trends in 2023.

The report states that the industry outpaced academia in AI development and deployment. Out of the 149 foundational models published in 2023, 108 (72.5%) were from industry compared to just 28 (18.8%) from academia.

Standford’s report reflects industry dominance and rising training costs in AI
Standford’s report reflects industry dominance and rising training costs in AI

Google (18) leads the way, followed by Meta (11), Microsoft (9), and OpenAI (7).

Standford’s report reflects industry dominance and rising training costs in AI
Standford’s report reflects industry dominance and rising training costs in AI

United States leads as the top source with 109 foundational models out of 149, followed by China (20) and the UK (9). In case of machine learning models, the United States again tops the chart with 61 notable models, followed by China (15) and France (8).

Regarding AI models’ training and computing costs, Gemini Ultra leads with a training cost of $191 million, followed by GPT-4, which has a training cost of $78 million.

Standford’s report reflects industry dominance and rising training costs in AI
Standford’s report reflects industry dominance and rising training costs in AI

Lastly, in 2023, AI reached human performance levels in many key AI benchmarks, such as reading comprehension, English understanding, visual thinking, image classification, etc.

Standford’s report reflects industry dominance and rising training costs in AI
Standford’s report reflects industry dominance and rising training costs in AI

Why does it matter?

Industry dominance in AI research suggests that companies will continue to drive advancements in the field, leading to more advanced and capable AI systems. However, the rising costs of AI training may pose challenges, as it could limit access to cutting-edge AI technology for smaller organizations or researchers.

Source

A Daily chronicle of AI Innovations April 16th 2024: 🎬 Adobe partners with OpenAI, RunwayML & Pika for Premiere Pro; 🚀 Reka launches Reka Core: their frontier in multimodal AI; 🇯🇵 OpenAI is opening its first international office in Tokyo;  🤖 Hugging Face has rolled out Idefics2 ; 💬 Quora’s Poe aims to become the ‘App Store’ for AI chatbots; 👥 Instagram is testing an AI program to amplify influencer engagement;  👩‍💻 Microsoft has released and open-sourced the new WizardLM-2 family of LLMs; 📋 Limitless AI launched a personal meeting assistant in a pendant 💰 Microsoft invests $1.5 billion in AI firm 📈 Baidu says its ChatGPT-like Ernie bot exceeds 200 million users 💻 OpenAI introduces Batch API with up to 50% discount for asynchronous tasks

Adobe partners with OpenAI, RunwayML & Pika for Premiere Pro

Adobe is integrating generative AI in Premiere Pro. The company is developing its own Firefly Video Model and teaming up with third-party AI models like OpenAI’s Sora, RunwayML, and Pika to bring features like Generative Extend, Object Addition and Removal, and Generative B-Roll to the editing timeline.

Adobe is committed to an open approach for delivering models. It allows editors to choose the best AI models for their needs to streamline video workflows, reduce tedious tasks, and expand creativity. It also provides “Content Credentials” to track model usage.

Why does this matter?

Adobe Premiere Pro has been used in blockbuster films like Deadpool, Gone Girl, and Everything Everywhere All at Once. By integrating generative AI into Premiere Pro, Adobe is transforming the film industry, allowing editors to streamline workflows and focus more on creative storytelling.

Source

Reka launches Reka Core: their frontier in multimodal AI

Another day, another GPT-4-class model. But this time, it’s not from the usual suspects like OpenAI, Google, or Anthropic. Reka, a lesser-known AI startup has launched a new flagship offering, Reka Core – a most advanced and one of only two commercially available comprehensive multimodal solutions. It excels at understanding images, videos, and audio while offering a massive context window, exceptional reasoning skills, and even coding.

Reka launches Reka Core: their frontier in multimodal AI
Reka launches Reka Core: their frontier in multimodal AI

It outperforms other models on various industry-accepted evaluation metrics. To provide flexibility, Reka Core can be deployed via API, on-premises, or on-device. Reka’s partnerships with Snowflake and Oracle are set to democratize access to this tech for AI innovation across industries.

Why does this matter?

Reka Core matches and even surpasses the performance of leading OpenAI, Google, and Anthropic models across various benchmarks and modalities. By offering cost-effective, multi-modal solutions, Reka has the potential to make advanced AI more accessible and drive new applications across multiple industries.

Source

OpenAI is opening its first international office in Tokyo

OpenAI is releasing a custom version of its GPT-4 model, specially optimized for the Japanese language. This specialized offering promises faster and more accurate performance and improved text handling.

Tadao Nagasaki has been appointed President of OpenAI Japan. The company plans to collaborate with the Japanese government, local businesses, and research institutions to develop safe AI tools that serve Japan’s unique needs. With Daikin and Rakuten already using ChatGPT Enterprise and local governments like Yokosuka City seeing productivity boosts, OpenAI is poised to impact the region significantly.

Why does it matter?

The move reflects OpenAI’s commitment to serving diverse markets. It could set a precedent for other AI companies, fostering a more inclusive and local approach. And as Japan grapples with rural depopulation and labor shortages, AI could prove invaluable in driving progress.

Source

💰 Microsoft invests $1.5 billion in AI firm

  • Microsoft will invest $1.5 billion in G42, a leading UAE artificial intelligence firm, as part of a strategic shift to align with American technology and disengage from Chinese partnerships following negotiations with the US government.
  • The investment enhances Microsoft’s influence in the Middle East, positioning G42 to use Microsoft Azure for its AI services, underpinning US efforts to limit Chinese access to advanced technologies.
  • This deal, which also involves Microsoft’s Brad Smith joining G42’s board, comes amidst broader US concerns about tech firms with Chinese ties.
  • Source

📈 Baidu says its ChatGPT-like Ernie bot exceeds 200 million users 

  • Baidu’s AI chatbot ‘Ernie Bot’ has reached 200 million users and its API is used 200 million times daily.
  • Ernie Bot, the first locally developed ChatGPT-like chatbot in China, was publicly released eight months ago after receiving approval from Beijing.
  • Despite its growth, Ernie Bot faces strong competition from rival domestic AI services, such as the Alibaba-backed ‘Kimi’ chatbot from Moonshot AI.
  • Source

💻 OpenAI introduces Batch API with up to 50% discount for asynchronous tasks

  • OpenAI introduces a new Batch API providing up to 50% discount for asynchronous tasks like summarization, translation, and image classification.
  • This Batch API allows results for bulk API requests within 24 hours by uploading a JSONL file of requests in batch format, currently supporting only the /v1/chat/completions endpoint.
  • OpenAI expects this to enable more efficient use of its APIs for applications that require a large number of requests.
  • Source

Unleash the Power of Generative AI: Build Breakthrough Apps with AWS Bedrock

Struggling to keep up with the rapid advancements in generative AI? AWS Bedrock offers a one-stop shop for developers. Access a variety of high-performing foundation models from leading names in AI, all through a single API. Fine-tune models with your data, leverage pre-built agents, and focus on building innovative applications.

For detailed study, refer the blog – https://www.seaflux.tech/blogs/aws-bedrock-models

What is AWS Bedrock?

AWS Bedrock is a fully managed service designed for developers to streamline the development of generative AI applications. It consists of high-performing foundation models (FMs) from leading AI companies that can be accessed from a single API. AWS Bedrock has tied its know with AI pros like AI21 Labs, Anthropic, Cohere, Meta, Stability AI, and also has in-house capabilities. Each FM has its own unique feature that can be leveraged according to your project preference. This eliminates the need for developers to manage the infrastructure and tooling necessary to train and deploy their own models. Despite the simplified process for developing applications, privacy and security aspects are not compromised. AWS Bedrock ensures the integrity and confidentiality of the developer’s data used for creating generative AI applications.

Key Features of AWS Bedrock

  • Variety of FMs: A wide range of high-performing models are available for different tasks like text generation, image generation, code generation, and more.
  • Simple API: A single API that makes it quick and easy to integrate FMs into your applications.
  • Fully managed service: All the infrastructure and tooling are managed for you to focus on building your applications.
  • Scalable: Applications can be scaled up or down as the requirement changes.
  • Secure: AWS Bedrock provides built-in security and privacy features ensuring integrity and confidentiality

How does AWS Bedrock work?

  1. Choose a foundation model: Browse the available models and select the one that best fits your needs.
  2. Send an API request: Use the simple API to send your data to the chosen model.
  3. Receive the output: The model will generate the desired output, such as text, code, or an image.
  4. Integrate the output: Use the output in your application however you like.

Types of Foundation Model (FM)

AWS Bedrocks provides 6 FMs with more than 15 versions that can be leveraged as per the project’s requirements. All the models are pre-trained on large datasets and are very reliable tools for a wide range of applications. The following table shows a brief about these FMs, and to know more about the models, visit the AWS Bedrock official website.
AAWS Bedrock, AWS Bedrock blog, aws solution to develop generative ai applications, Token size of AWS bedrock models, Languages supported by aws bedrock foundation models, Different use cases of AWS bedrock foundation models

AWS Bedrock Pricing

AWS Bedrock provides two types of pricing models and charges based on the model inference and customization in the model.

  1. On-demand & Batch: The pay-as-you-go pricing model is used without any time-based commitments.
  2. Provisioned Throughput: A sufficient throughput is provided in exchange for a time-based commitment to meet the performance demand of the application. The term can be 1 month or 6-month commitment.
Follow through this pricing table to compare the models, or visit the AWS Bedrock official pricing website to know more about it.
AWS bedrock pricing, aws solution to develop generative ai applications, AI21 labs pricing, Anthropic claude pricing, Cohere Command pricing, Stabality AI XL pricing

What Else Is Happening in AI on April 16th, 2024❗

🤖 Hugging Face has rolled out Idefics2 

Hugging Face has released Idefics2, a more compact and capable version of its visual language model. With just 8 billion parameters, this open-source model enhances image manipulation, improves OCR, and answers questions on visual data. (Link)

💬 Quora’s Poe aims to become the ‘App Store’ for AI chatbots

After a $75 million funding round, Poe has launched a “multi-bot chat” feature that allows users to seamlessly integrate various AI models into a single conversation. Positioning itself as the “app store” for chatbots, Poe is also rolling out monetization tools for creators and planning an enterprise tier for businesses. (Link)

👥 Instagram is testing an AI program to amplify influencer engagement

The “Creator AI” program lets popular creators interact with fans through automated chatbots. The bots will mimic the influencer’s voice using their past content, aiming to boost engagement while cutting down on manual responses. While some creators worry this could undermine authenticity, Meta sees AI as crucial to its future. (Link)

👩‍💻 Microsoft has released and open-sourced the new WizardLM-2 family of LLMs

This next-gen LLM lineup boasts three cutting-edge versions—The 8x22B model outperforms even the best open-source alternatives, while the 70B and 7B variants deliver best-in-class reasoning and efficiency, respectively. (Link)

📋 Limitless AI launched a personal meeting assistant in a pendant

Limitless launched a $99 wearable “Limitless Pendant” to transcribe conversations, generate real-time notes, and seamlessly integrate with your work apps. While starting with a focus on meetings, the startup’s CEO Dan Siroker sees Limitless eventually doing much more – proactively surfacing relevant information and even automating tasks on your behalf. (Link)

A Daily chronicle of AI Innovations April 15th 2024: 🚗 Tesla lays off more than 10% of its workforce 🎥 Adobe explores OpenAI partnership as it adds AI video tools 📱 Apple’s AI features on iOS 18 may run locally on your iPhone 📊 xAI’s first multimodal model with a unique dataset ♾️ Infini-Attention: Google’s breakthrough gives LLMs limitless context ⚠️ Adobe’s Firefly AI trained on competitor’s images: Bloomberg report

xAI’s first multimodal model with a unique dataset

xAI, Elon Musk’s AI startup, has released the preview of Grok-1.5V, its first-generation multimodal AI model. This new model combines strong language understanding capabilities with the ability to process various types of visual information, like documents, diagrams, charts, screenshots, and photographs.

The startup claims Grok-1.5V has shown competitive performance across several benchmarks, including tests for multidisciplinary reasoning, mathematical problem-solving, and visual question answering. One notable achievement is its exceptional performance on the RealWorldQA dataset, which evaluates real-world spatial understanding in AI models.

Developed by xAI, this dataset features over 700 anonymized images from real-world scenarios, each accompanied by a question and verifiable answer. The release of Grok-1.5V and the RealWorldQA dataset aims to advance the development of AI models that can effectively comprehend and interact with the physical world.

Why does this matter?

What makes Grok-1.5V unique is its integration with the RealWorldQA dataset, which focuses on real-world spatial understanding crucial for AI systems in physical environments. The public availability of this dataset could significantly advance the development of AI-driven robotics and autonomous systems. With Musk’s backing, xAI could lead in multimodal AI and contribute to reshaping human-AI interaction.

Source

Infini-Attention: Google’s breakthrough gives LLMs limitless context

Google researchers have developed a new technique called Infini-attention that allows LLMs to process text sequences of unlimited length. By elegantly modifying the Transformer architecture, Infini-attention enables LLMs to maintain strong performance on input sequences exceeding 1 million tokens without requiring additional memory or causing exponential increases in computation time.

Infini-Attention: Google's breakthrough gives LLMs limitless context
Infini-Attention: Google’s breakthrough gives LLMs limitless context

The key innovation behind Infini-attention is the addition of a “compressive memory” module that efficiently stores old attention states once the input sequence grows beyond the model’s base context length. This compressed long-range context is then aggregated with local attention to generate coherent and contextually relevant outputs.

In benchmark tests on long-context language modeling, summarization, and information retrieval tasks, Infini-attention models significantly outperformed other state-of-the-art long-context approaches while using up to 114 times less memory.

Why does this matter?

Infini-attention can help AI systems expertly organize, summarize, and surface relevant information from vast knowledge bases. Additionally, infinite contextual understanding can help AI systems generate more nuanced and contextually relevant long-form content like articles, reports, and creative writing pieces. Overall, we can expect AI tools to generate more valuable and less generic content with this technique.

Source

Adobe’s Firefly AI trained on competitor’s images: Bloomberg report

In a surprising revelation, Adobe’s AI image generator Firefly was found to have been trained not just on Adobe’s own stock photos but also on AI-generated images from rival platforms like Midjourney and DALL-E. The Bloomberg report, which cites insider sources, notes that while these AI images made up only 5% of Firefly’s training data, their inclusion has sparked an internal ethics debate within Adobe.

The news is particularly noteworthy given Adobe’s public emphasis on Firefly’s “ethical” sourcing of training data, a stance that aimed to differentiate it from competitors. The company had even set up a bonus scheme to compensate artists whose work was used to train Firefly. However, the decision to include AI-generated images, even if labeled as such by the submitting artists, has raised questions about the consistency of Adobe’s ethical AI practices.

Why does it matter?

As AI systems learn from one another in a continuous feedback loop, the distinction between original creation, inspiration, and imitation becomes blurred. This raises complex issues around intellectual property rights, consent, and the difference between remixing and replicating. Moreover, the increasing prevalence of AI-generated content in training data sets could lead to a homogenization of AI outputs, potentially stifling creativity and diversity.

Source

🚗 Tesla lays off more than 10% of its workforce 

  • Tesla plans to lay off “more than 10 percent” of its global workforce following its first year-over-year decline in vehicle deliveries since 2020, impacting at least 14,000 employees.
  • CEO Elon Musk expressed regret over the layoffs in an internal email, stating they are necessary for the company to remain “lean, innovative and hungry” for future growth.
  • Senior vice president Drew Baglino and policy chair Rohan Patel are among the top executives reported to be leaving the company amid these changes.
  • Source

🎥 Adobe explores OpenAI partnership as it adds AI video tools 

  • Adobe is enhancing Premiere Pro with new AI video tools, enabling capabilities such as video generation, object addition/removal, and clip extension, and is exploring a potential partnership with OpenAI.
  • The integration of OpenAI’s Sora with Adobe’s video tools is considered an “early exploration,” aiming to augment Adobe’s offerings and provide users with advanced generative capabilities.
  • Adobe aims to offer more choice to Premiere Pro users by potentially integrating third-party AI models and adding Content Credentials to identify the AI used, despite current limitations and the unclear extent of user control over these new features.
  • Source

📱 Apple’s AI features on iOS 18 may run locally on your iPhone LINK

  • Apple’s iOS 18, set to debut at WWDC 2024 on June 10, promises to be the most significant software upgrade with enhanced features like a smarter Siri through generative AI.
  • According to Bloomberg’s Mark Gurman, the initial set of AI features in iOS 18 will operate entirely on-device without requiring cloud processing, ensuring privacy and efficiency.
  • Apple is in discussions with AI developers such as Google’s Gemini, OpenAI’s GPT, and Baidu to integrate generative AI tools into iOS 18, potentially including third-party AI chatbots.

What Else Is Happening in AI on April 15th 2024❗

🤖 Meta trials AI chatbot on WhatsApp, Instagram, and Messenger

Meta is testing its AI chatbot, Meta AI, with WhatsApp, Instagram, and Messenger users in India and parts of Africa. The move allows Meta to leverage its massive user base across these apps to scale its AI offerings. Meta AI can answer user queries, generate images from text prompts, and assist with Instagram search queries. (Link)

🎨 Ideogram introduces new features to its AI image generation model

Ideogram’s AI image generation model now offers enhanced capabilities like description-based referencing, negative prompting, and options for generating images at varying speeds and quality levels. The upgrade aims to improve image coherence, photorealism, and text rendering quality, with human raters showing a 30-50% preference for the new version over the previous one. (Link)

🖼️ New Freepik AI tool redefines image generation with realism and versatility

Freepik has launched the latest version of its AI Image Generator that offers real-time generation, infinite variations, and photorealistic results. The tool allows users to create infinite variations of an image with intuitive prompts, combining colors, settings, characters, and scenarios. It delivers highly realistic results and offers a streamlined workflow with real-time generation and infinite scrolling. (Link)

💼 OpenAI promoted ChatGPT Enterprise to corporates with road-show-like events

OpenAI CEO Sam Altman recently hosted events in San Francisco, New York, and London, pitching ChatGPT Enterprise and other AI services to hundreds of Fortune 500 executives. This move is part of OpenAI’s strategy to diversify revenue streams and compete with partner Microsoft in selling AI products to enterprises. The events showcased applications such as call center management, translation, and custom AI solutions. (Link)

📔 Google’s Notes tool now offers custom AI-generated backgrounds

Google has introduced an AI-powered background generation feature for its experimental Notes tool, allowing users to personalize their notes with custom images created from text prompts. The feature, currently available for select users in the US and India, utilizes Google’s Gemini AI model for image generation. (Link)

A Daily chronicle of AI Innovations April 12th 2024: 💥 OpenAI fires two researchers for alleged leaking; 🍎 Apple is planning to bring new AI-focused M4 chips to entire line of Macs; 🤷‍♀️ Amazon CEO: don’t wait for us to launch a ChatGPT competitor; 💬 ChatGPT GPT-4 just got a huge upgrade; 🧠 Gabe Newell, the man behind Steam, is working on a brain-computer interface; 🔍 Cohere’s Rerank 3 powers smarter enterprise search; 💻 Apple M4 Macs: Coming soon with AI power!; 📝 Meta’s OpenEQA puts AI’s real-world comprehension to test

Cohere’s Rerank 3 powers smarter enterprise search

Cohere has released a new model, Rerank 3, designed to improve enterprise search and Retrieval Augmented Generation (RAG) systems. It can be integrated with any database or search index and works with existing legacy applications.

Cohere’s Rerank 3 powers smarter enterprise search
Cohere’s Rerank 3 powers smarter enterprise search

Rerank 3 offers several improvements over previous models:

  • It handles a longer context of documents (up to 4x longer) to improve search accuracy, especially for complex documents.
  • Rerank 3 supports over 100 languages, addressing the challenge of multilingual data retrieval.
  • The model can search various data formats like emails, invoices, JSON documents, codes, and tables.
  • Rerank 3 works even faster than previous models, especially with longer documents.
  • When used with Cohere’s RAG systems, Rerank 3 reduces the cost by requiring fewer documents to be processed by the expensive LLMs.

Plus, enterprises can access it through Cohere’s hosted API, AWS Sagemaker, and Elasticsearch’s inference API.

Why does this matter?

Rerank 3 represents a step towards a future where data is not just stored but actively used by businesses to make smarter choices and automate tasks. Imagine instantly finding a specific line of code from an email or uncovering pricing details buried in years of correspondence.

Source

Apple M4 Macs: Coming soon with AI power!

Apple is overhauling its Mac lineup with a new M4 chip focused on AI processing. This comes after the recent launch of M3 Macs, possibly due to slowing Mac sales and similar features in competitor PCs.

The M4 chip will come in three tiers (Donan, Brava, Hidra) and will be rolled out across various Mac models throughout 2024 and early 2025. Lower-tier models like MacBook Air and Mac Mini will get the base Donan chip, while high-performance Mac Pro will be equipped with the top-tier Hidra. We can expect to learn more about the specific AI features of the M4 chip at Apple’s WWDC on June 10th.

Why does this matter?

Apple’s new AI-powered M4 Mac chip could make Macs much faster for things like video editing and scientific work, competing better with computers with similar AI features.

By controlling hardware and software, Apple can fine-tune everything to ensure a smooth user experience and future improvements.

Source

Meta’s OpenEQA puts AI’s real-world comprehension to test

Meta AI has released a new dataset called OpenEQA to measure how well AI understands the real world. This “embodied question answering” (EQA) involves an AI system being able to answer questions about its environment in natural language.

The dataset includes over 1,600 questions about various real-world places and tests an AI’s ability to recognize objects, reason about space and function, and use common sense knowledge.

Why does this matter?

While OpenEQA challenges AI with questions demanding visual and spatial reasoning, it also exposes limitations in current AI models that often rely solely on text knowledge. Its role could push researchers to develop AI with a stronger grasp of the physical world.

Source

💥 OpenAI fires two researchers for alleged leaking

  • OpenAI has dismissed two researchers, Leopold Aschenbrenner and Pavel Izmailov, for allegedly leaking information following an undisclosed internal investigation.
  • The leaked information may be related to a research project called Q*, which involved a breakthrough in AI models solving unseen math problems, raising concerns about the lack of safeguards for commercializing such advanced technology.
  • The firings highlight a potential contradiction in OpenAI’s mission, as the company faces criticism for moving away from its original ethos of openness and transparency.
  • Source

🍎 Apple is planning to bring new AI-focused M4 chips to entire line of Macs

  • Apple is poised to launch its next-generation M4 chips as early as this year, aimed at enhancing AI capabilities and rejuvenating Mac sales following a 27% drop last fiscal year.
  • The M4 chips, reported to be nearing production, are expected to come in three variants named Donan, Brava, and Hidra, supporting a range of Mac products, including updates to the iMac, MacBook Pros, and Mac Mini initially, with the MacBook Air and Mac Studio to follow.
  • This accelerated update cycle to introduce M4 chips may lead to a short lifespan for the recently launched M3 chips, indicating Apple’s urgency to compete in the AI technology space against rivals with similar AI-focused hardware advancements.
  • Source

🤷‍♀️ Amazon CEO: don’t wait for us to launch a ChatGPT competitor

  • Amazon CEO Andy Jassy emphasizes the company’s focus on building foundational “primitives” for generative AI rather than quickly launching public-facing products like a ChatGPT competitor.
  • Amazon has launched AI products such as Amazon Bedrock and Amazon Q aimed at software engineers and business customers, aligning with its strategy to empower third-party developers to create GenAI applications.
  • Despite not directly competing with ChatGPT, Amazon is investing in the AI domain, including a $4 billion investment in AI company Anthropic, while also enhancing its existing products like Alexa with AI capabilities.
  • Source

💬 ChatGPT GPT-4 just got a huge upgrade 

  • ChatGPT’s GPT-4 Turbo model has received an upgrade, enhancing its abilities in writing, math, logical reasoning, and coding, as announced by OpenAI for its premium users.
  • The upgrade, distinguished by significant performance improvements in mathematics and GPQA, also aims for more succinct, direct, and conversational responses.
  • This new version of ChatGPT, which includes data up until December 2023, shows improved performance on recent topics, such as acknowledging the launch of the iPhone 15.
  • Source

🧠 Gabe Newell, the man behind Steam, is working on a brain-computer interface

  • Gabe Newell, co-founder of Valve and the force behind Steam, has been developing a brain-computer interface (BCI) technology through a venture named Starfish Neuroscience, rivaling Elon Musk’s Neuralink.
  • Since 2019, Newell has explored gaming applications for BCIs and discussed potential future capabilities like editing feelings, highlighting the technology’s potential beyond traditional interfaces.
  • Aside from his BCI pursuits, Newell has faced recent challenges including an antitrust lawsuit against Steam and the sale of his megayacht, amidst managing COVID-19 precautions and legal appearances.
  • Source

What Else Is Happening in AI on April 12th 2024❗

🔄 ChatGPT gets an upgrade for premium users

OpenAI has released an enhanced version of GPT-4 Turbo for ChatGPT Plus, Team, and Enterprise customers. The new model, trained on data until December 2023, promises more direct responses, less verbosity, and improved conversational language, along with advancements in writing, math, reasoning, and coding. (Link)

🤝 Dr. Andrew Ng joins Amazon’s Board of Directors

Amazon has appointed Dr. Andrew Ng, a renowned AI expert and founder of several influential AI companies, to its Board of Directors. With his deep expertise in machine learning and AI education, Ng is expected to provide valuable insights as Amazon navigates the transformative potential of generative AI. (Link)

⌚️ Humane’s $699 Ai Pin hits the US market

Humane’s Ai Pin is now available across the US, with global expansion on the horizon through SKT and SoftBank partnerships. The wearable AI device is powered by a $24/month plan, including unlimited AI queries, data, and storage. The international availability is to be announced soon. (Link)

📱 TikTok might use AI influencers for ads

TikTok is developing a new feature that lets companies use AI characters to advertise products. These AI influencers can read scripts made by advertisers or sellers. TikTok has been testing this feature but isn’t sure when it will be available for everyone to use. (Link)

🤖 Sanctuary AI’s humanoid robot to be tested at Magna

Magna, a major European car manufacturer, will pilot Sanctuary AI’s humanoid robot, Phoenix, at one of its facilities. This follows similar moves by other automakers exploring the use of humanoid robots in manufacturing, as companies seek to determine the potential return on investment. (Link)

A Daily chronicle of AI Innovations April 11th 2024: 🚀 Meta unveils next-generation AI chip for enhanced workloads 🎶 New AI tool lets you generate 1200 songs per month for free 💰 Adobe is buying videos for $3 per minute to build an AI model 🤖 Google expands Gemma family with new models 🌐 Mistral unveils Mixtral-8x22B open language model 📷 Google Photos introduces free AI-powered editing tools 🖼️ Microsoft enhances Bing visual search with personalization 🛡️ Sama red team: Safety-centered solution for Generative AI 💥 Apple hit with ‘mercenary spyware attacks’  🧠 Humane AI has only one problem: it just doesn’t work 🔍 MistralAI unveils groundbreaking open model Mixtral 8x22B 🙃 Microsoft proposed using DALL-E to US military last year 🎵 New AI music generator Udio synthesizes realistic music on demand 🎬 Adobe is purchasing video content to train its AI model

🚀 Meta unveils next-generation AI chip for enhanced workloads

Meta has introduced the next generation of its Meta Training and Inference Accelerator (MTIA), significantly improving on MTIAv1 (its first-gen AI inference accelerator). This version more than doubles the memory and compute bandwidth, designed to effectively serve Meta’s crucial AI workloads, such as its ranking and recommendation models and Gen AI workloads.

Meta has also co-designed the hardware system, the software stack, and the silicon, which is essential for the success of the overall inference solution.

Meta unveils next-generation AI chip for enhanced workloads
Meta unveils next-generation AI chip for enhanced workloads

Early results show that this next-generation silicon has improved performance by 3x over the first-generation chip across four key models evaluated. MTIA has been deployed in the data center and is now serving models in production.

Why does this matter?

This is a bold step towards self-reliance in AI! Because Meta controls the whole stack, it can achieve an optimal mix of performance and efficiency on its workloads compared to commercially available GPUs. This eases NVIDIA’s grip on it, which might be having a tough week with other releases, including Intel’s Gaudi 3 and Google Axion Processors.

Source

New AI tool lets you generate 1200 songs per month for free

Udio, a new AI music generator created by former Google DeepMind researchers, is now available in beta. It allows users to generate up to 1200 songs per month for free, with the ability to specify genres and styles through text prompts.

The startup claims its AI can produce everything from pop and rap to gospel and blues, including vocals. While the free beta offers limited features, Udio promises improvements like longer samples, more languages, and greater control options in the future. The company is backed by celebrities like Will.i.am and investors like Andreessen Horowitz.

Why does this matter?

AI-generated music platforms like Udio democratize music creation by making it accessible to everyone, fostering new artists and diverse creative expression. This innovation could disrupt traditional methods, empowering independent creators lacking access to expensive studios or musicians.

Source

💰 Adobe is buying videos for $3 per minute to build an AI model

Adobe is buying videos at $3 per minute from its network of photographers and artists to build a text-to-video AI model. It has requested short clips of people engaged in everyday actions such as walking or expressing emotions including joy and anger, interacting with objects such as smartphones or fitness equipment, etc.

The move shows Adobe trying to catch up to competitors like OpenAI (Sora). Over the past year, Adobe has added generative AI features to its portfolio, including Photoshop and Illustrator, that have garnered billions of uses. However, Adobe may be lagging behind the AI race and is trying to catch up.

Why does this matter?

Adobe’s targeted video buying for AI training exposes the hefty price tag of building competitive AI. Smaller companies face an uphill battle—they might need to get scrappier, focus on specific niches, team up, or use free, open-source AI resources.

Source

💥 Apple hit with ‘mercenary spyware attacks

  • Apple has issued a warning to iPhone users in 92 countries about a potential “mercenary spyware attack” aimed at compromising their devices, without identifying the attackers or the consequences.
  • The company suggests that the attack is highly targeted, advising recipients to take the warning seriously and to update their devices with the latest security patches and practice strong cyber hygiene.
  • This type of attack is often linked to state actors employing malware from private companies, with the infamous ‘Pegasus’ spyware mentioned as an example, capable of extensive surveillance on infected phones.
  • Source

🧠 Humane AI has only one problem: it just doesn’t work

  • The Humane AI Pin, retailing for $699 plus a $24 monthly fee, is designed as a wearable alternative to smartphones, promising users freedom from their screens through AI-assisted tasks. However, its functionality falls significantly short of expectations.
  • Throughout testing, the AI Pin struggled with basic requests and operations, demonstrating unreliability and slow processing times, leading to the conclusion that it fails to deliver on its core promise of a seamless, smartphone-free experience.
  • Despite its well-intentioned vision for a post-smartphone future and the integration of innovative features like a screenless interface and ambient computing, the device’s current state of performance and high cost make it a poor investment for consumers.
  • Source

🔍 MistralAI unveils groundbreaking open model Mixtral 8x22B

  • Mistral AI has released Mixtral 8x22B, an open-source AI model boasting 176 billion parameters and a 65,000-token context window, expected to surpass its predecessor and compete with major models like GPT-3.5 and Llama 2.
  • The Paris-based startup, valued at over $2 billion, aims to democratize access to cutting-edge AI by making Mixtral 8x22B available on platforms like Hugging Face and Together AI, allowing for widespread use and customization.
  • Despite its potential for innovation in fields like customer service and drug discovery, Mixtral 8x22B faces challenges related to its “frontier model” status, including the risk of misuse due to its open-source nature and lack of control over harmful applications.
  • Source

🙃 Microsoft proposed using DALL-E to US military last year

  • Microsoft proposed to the U.S. Department of Defense in 2023 to use OpenAI’s DALL-E AI for software development in military operations.
  • The proposal included using OpenAI tools like ChatGPT and DALL-E for document analysis, machine maintenance, and potentially training battlefield management systems with synthetic data.
  • Microsoft had not implemented the use of DALL-E in military projects, and OpenAI, which did not participate in Microsoft’s presentation, restricts its technology from being used to develop weapons or harm humans.
  • Source

🎵 New AI music generator Udio synthesizes realistic music on demand

  • Uncharted Labs has officially launched its music generator, Udio, which can transform text prompts into professional-quality music tracks, challenging the leading AI music generator, Suno V3.
  • Udio has impressed users and reviewers alike with its ability to generate songs that feature coherent lyrics, well-structured compositions, and competitive rhythms, some even considering it superior to Suno V3.
  • Despite facing initial server overload due to high user demand, Udio’s user-friendly interface and strong backing from notable investors suggest a promising future for AI-assisted music creation, though it remains free during its beta testing phase.
  • Source

🎬 Adobe is purchasing video content to train its AI model

  • Adobe is developing a text-to-video AI model, offering artists around $3 per minute for video footage to train the new tool, as reported by Bloomberg.
  • The software company has requested over 100 video clips from artists, aiming for content that showcases various emotions and activities, but has set a low budget for acquisitions.
  • Despite the potential for AI to impact artists’ future job opportunities and the lack of credit or royalties for the contributed footage, Adobe is pushing forward with the AI model development.
  • Source

What Else Is Happening in AI on April 11th 2024❗

🤖 Google expands Gemma family with new models

Google has expanded its Gemma family with two new models: CodeGemma and RecurrentGemma. CodeGemma is tailored for developers, offering intelligent code completion and chat capabilities for languages like Python and JavaScript. RecurrentGemma is optimized for efficiency in research, utilizing recurrent neural networks and local attention. (Link)

🌐 Mistral unveils Mixtral-8x22B open language model

Mistral AI has unveiled Mixtral-8x22B, a new open language model with extensive capabilities. This model, featuring 64,000 token context windows and requiring 258GB of VRAM, is a mixture-of-experts model. Early users are exploring its potential, with more details expected soon. (Link)

📷 Google Photos introduces free AI-powered editing tools

Google Photos is rolling out free AI-powered editing tools for all users starting May 15. Features like Magic Eraser, Photo Unblur, and Portrait Light will be accessible without a subscription. Pixel users will also benefit from the Magic Editor, which simplifies complex edits using generative AI. (Link)

🖼️ Microsoft enhances Bing visual search with personalization

Microsoft enhances Bing Visual Search with personalized visual systems based on user preferences. A patent application reveals that search results will be tailored to individual interests, such as showing gardening-related images to gardening enthusiasts and food-related visuals to chefs. (Link)

🛡️ Sama red team: Safety-centered solution for Generative AI

Sama has introduced Sama Red Team, a safety-centered solution for evaluating risks associated with generative AI and LLMs. This system simulates adversarial attacks to identify vulnerabilities related to bias, personal information, and offensive content, contributing to a more ethical AI landscape. (Link)

A Daily chronicle of AI Innovations April 10th 2024: 👀 OpenAI gives GPT-4 a major upgrade; 💬 Quora’s Poe now lets AI chatbot developers charge per message; 🌐 Google updates and expands its open source Gemma AI model family; 🔥 Intel unveils latest AI chip as Nvidia competition heats up; 📱 WordPress parent acquires Beeper app which brought iMessage to Android; 🤔 New bill would force AI companies to reveal use of copyrighted art; 🧠 Intel’s new AI chip: 50% faster, cheaper than NVIDIA’s; 🤖 Meta to Release Llama 3 Open-source LLM next week; ☁️ Google Cloud announces major updates to enhance Vertex AI

Intel’s new AI chip: 50% faster, cheaper than NVIDIA’s

Intel has unveiled its new Gaudi 3 AI accelerator, which aims to compete with NVIDIA’s GPUs. According to Intel, the Gaudi 3 is expected to reduce training time for large language models like Llama2 and GPT-3 by around 50% compared to NVIDIA’s H100 GPU. The Gaudi 3 is also projected to outperform the H100 and H200 GPUs in terms of inference throughput, with around 50% and 30% faster performance, respectively.

Intel's new AI chip: 50% faster, cheaper than NVIDIA's
Intel’s new AI chip: 50% faster, cheaper than NVIDIA’s

The Gaudi 3 is built on a 5nm process and offers several improvements over its predecessor, including doubling the FP8, quadrupling the BF16 processing power, and increasing network and memory bandwidth. Intel is positioning the Gaudi 3 as an open, cost-effective alternative to NVIDIA’s GPUs, with plans to make it available to major OEMs starting in the second quarter of 2024. The company is also working to create an open platform for enterprise AI with partners like SAP, Red Hat, and VMware.

Why does it matter?

Intel is challenging NVIDIA’s dominance in the AI accelerator market. It will introduce more choice and competition in the market for high-performance AI hardware. It could drive down prices, spur innovation, and give customers more flexibility in building AI systems. The open approach with community-based software and standard networking aligns with broader trends toward open and interoperable AI infrastructure.

Source

Meta to release Llama 3 open-source LLM next week

Meta plans to release two smaller versions of its upcoming Llama 3 open-source language model next week. These smaller models will build anticipation for the larger version, which will be released this summer. Llama 3 will significantly upgrade over previous versions, with about 140 billion parameters compared to 70 billion for the biggest Llama 2 model. It will also be a more capable, multimodal model that can generate text and images and answer questions about images.

The two smaller versions of Llama 3 will focus on text generation. They’re intended to resolve safety issues before the full multimodal release. Previous Llama models were criticized as too limited, so Meta has been working to make Llama 3 more open to controversial topics while maintaining safeguards.

Why does it matter?

The open-source AI model landscape has become much more competitive in recent months, with other companies like Mistral and Google DeepMind also releasing their own open-source models. Meta hopes that by making Llama 3 more open and responsive to controversial topics, it can catch up to models like OpenAI’s GPT-4 and become a standard for many AI applications.

Source

Google Cloud announces major updates to enhance Vertex AI

Google Cloud has announced exciting model updates and platform capabilities that continue to enhance Vertex AI:

  • Gemini 1.5 Pro: Gemini 1.5 Pro is now available in public preview in Vertex AI, the world’s first one million-token context window to customers. It also supports the ability to process audio streams, including speech and even the audio portion of videos.
  • Imagen 2.0: Imagen 2.0 can now create short, 4-second live images from text prompts, enabling marketing and creative teams to generate animated content. It also has new image editing features like inpainting, outpainting, and digital watermarking.
  • Gemma: Google Cloud is adding CodeGemma to Vertex AI. CodeGemma is a new lightweight model from Google’s Gemma family based on the same research and technology used to create Gemini.
  • MLOps: To help customers manage and deploy these large language models at scale, Google has expanded the MLOps capabilities for Gen AI in Vertex AI. This includes new prompt management tools for experimenting, versioning, optimizing prompts, and enhancing evaluation services to compare model performance.

Why does it matter?

These updates significantly enhance Google Cloud’s generative AI offerings. It also strengthens Google’s position in the generative AI space and its ability to support enterprise adoption of these technologies.

Source

👀 OpenAI gives GPT-4 a major upgrade

  • OpenAI has introduced GPT-4 Turbo with Vision, a new model available to developers that combines text and image processing capabilities, enhancing AI chatbots and other applications.
  • This multimodal model, which maintains a 128,000-token window and knowledge from December 2023, simplifies development by allowing a single model to understand both text and images.
  • GPT-4 Turbo with Vision simplifies development processes for apps requiring multimodal inputs like coding assistance, nutritional insights, and website creation from drawings.
  • Source

💬 Quora’s Poe now lets AI chatbot developers charge per message

  • Poe, a Quora-owned AI chatbot platform, introduced a new revenue model allowing creators to earn money by setting a price-per-message for their bots.
  • The revenue model aims to compensate creators for operational costs, fostering a diverse ecosystem of bots ranging from tutoring to storytelling.
  • This monetization strategy is initially available to U.S. creators, complemented by an analytics dashboard to track earnings and bot usage.
  • Source

🌐 Google updates and expands its open source Gemma AI model family

  • Google has enhanced the Gemma AI model family with new code completion models and improvements for more efficient inference, along with more flexible terms of use.
  • Three new versions of CodeGemma have been introduced, including a 7 billion parameter model for code generation and discussion, and a 2 billion parameter model optimized for fast code completion on local devices.
  • Google also unveiled RecurrentGemma, a model leveraging recurrent neural networks for better memory efficiency and speed in text generation, indicating a shift towards optimizing AI performance on devices with limited resources.
  • Source

🔥 Intel unveils latest AI chip as Nvidia competition heats up

  • Intel introduced its latest artificial intelligence chip, Gaudi 3, highlighting its efficiency and speed advantages over Nvidia’s H100 GPU and offering configurations that enhance AI model training and deployment.
  • The Gaudi 3 chip, which outperforms Nvidia in power efficiency and AI model processing speed, will be available in the third quarter, with Dell, Hewlett Packard Enterprise, and Supermicro among the companies integrating it into their systems.
  • Despite Nvidia’s dominant position in the AI chip market, Intel is seeking to compete by emphasizing Gaudi 3’s competitive pricing, open network architecture, and partnerships for open software development with companies like Google, Qualcomm, and Arm.
  • Source

📱 WordPress parent acquires Beeper app which brought iMessage to Android

  • Automattic, the owner of WordPress and Tumblr, has acquired Beeper, a startup known for its Beeper Mini app that attempted to challenge Apple’s iMessage, for $125 million despite the app’s quick defeat.
  • Beeper CEO Eric Migicovsky will oversee the merging of Beeper with Automattic’s similar app Texts, aiming to create the best chat app, with the combined service expected to launch later this year.
  • The acquisition raises questions due to Beeper Mini’s brief success and upcoming changes like Apple introducing RCS support to iPhones, but Automattic sees potential in Beeper’s stance on open messaging standards and its established brand.
  • Source

🤔 New bill would force AI companies to reveal use of copyrighted art

  • A new bill introduced in the US Congress by Congressman Adam Schiff aims to make artificial intelligence companies disclose the copyrighted material used in their generative AI models.
  • The proposed Generative AI Copyright Disclosure Act would require AI companies to register copyrighted works in their training datasets with the Register of Copyrights before launching new AI systems.
  • The bill responds to concerns about AI firms potentially using copyrighted content without permission, amidst growing litigation and calls for more regulation from the entertainment industry and artists.
  • Source

What Else Is Happening in AI on April 10th 2024❗

🚀 OpenAI launches GPT-4 Turbo with Vision model through API

OpenAI has unveiled the latest addition to its AI arsenal, the GPT -4 Turbo with Vision model, which is now “generally available” through its API. This new version has enhanced capabilities, including support for JSON mode and function calling for Vision requests. The upgraded GPT-4 Turbo model promises improved performance and is set to roll out in ChatGPT.  (Link)

👂 Google’s Gemini 1.5 Pro can now listen to audio

Google’s update to Gemini 1.5 Pro gives the model ears. It can process text, code, video, and uploaded audio streams, including audio from video, which it can listen to, analyze, and extract information from without a corresponding written transcript.(Link)

💰 Microsoft to invest $2.9 billion in Japan’s AI and cloud infrastructure

Microsoft announced it would invest $$2.9 billion over the next two years to increase its hyperscale cloud computing and AI infrastructure in Japan. It will also expand its digital skilling programs with the goal of providing AI skills to more than 3 million people over the next three years. (Link)

👩‍💻 Google launches Gemini Code Assist, the latest challenger to GitHub’s Copilot

At its Cloud Next conference, Google unveiled Gemini Code Assist, its enterprise-focused AI code completion and assistance tool. It provides various functions such as enhanced code completion, customization, support for various repositories, and integration with Stack Overflow and Datadog. (Link)

🛍️ eBay launches AI-driven ‘Shop the Look’ feature on its iOS app

eBay launched an AI-powered feature to appeal to fashion enthusiasts – “Shop the Look” on its iOS mobile application. It will suggest a carousel of images and ideas based on the customer’s shopping history. The recommendations will be personalized to the end user. The idea is to introduce how other fashion items may complement their current wardrobe. (Link)

A Daily chronicle of AI Innovations April 09th 2024: 🤖 Stability AI launches multilingual Stable LM 2 12B 📱 Ferret-UI beats GPT-4V in mobile UI tasks ⏰ Musk says AI will outsmart humans within a year 🍁 Canada bets big on AI with $2.4B investment 🎥 OpenAI is using YouTube for GPT-4 training 🤖 Meta to launch new Llama 3 models 👂 Google’s Gemini 1.5 Pro can now hear 💥 Google’s first Arm-based CPU will challenge Microsoft and Amazon in the AI race 📈 Boosted by AI, global PC market bounces back

🤖 Meta to launch new Llama 3 models

  • According to an insider, Meta will release two smaller versions of its planned major language model, Llama 3, next week to build anticipation for the major release scheduled for this summer.
  • The upcoming Llama 3 model, which will include both text generation and multimodal capabilities, aims to compete with OpenAI’s GPT-4 and is reported to potentially have up to 140 billion parameters.
  • Meta’s investment in the Llama 3 model and open-source AI reflects a broader trend of tech companies leveraging these technologies to set industry standards, similar to Google’s strategy with Android.
  • Source

👂 Google’s Gemini 1.5 Pro can now hear

  • Google has enhanced Gemini 1.5 Pro to interpret audio inputs, allowing it to process information from sources like earnings calls or video audio directly without needing a transcript.
  • Gemini 1.5 Pro, positioned as a mid-tier option within the Gemini series, now outperforms even the more advanced Gemini Ultra by offering faster and more intuitive responses without requiring model fine-tuning.
  • Alongside Gemini 1.5 Pro updates, Google introduced enhancements to its Imagen 2 model, including inpainting and outpainting features, and debuted a digital watermarking technology, SynthID, for tracking the origin of generated images.
  • Source

💥 Google’s first Arm-based CPU will challenge Microsoft and Amazon in the AI race

  • Google is developing its own Arm-based CPU named Axion to enhance AI operations in data centers and will launch it for Google Cloud business customers later this year.
  • The Axion CPU will improve performance by 30% over general-purpose Arm chips and by 50% over Intel’s processors, and it will support services like Google Compute Engine and Google Kubernetes Engine.
  • Google’s move to create its own Arm-based CPU and update its TPU AI chips aims to compete with Microsoft and Amazon in the AI space and reduce reliance on external suppliers like Intel and Nvidia.
  • Source

📈 Boosted by AI, global PC market bounces back

  • The global PC market has seen growth for the first time in over two years, with a 1.5% increase in shipments to 59.8 million units in the first quarter, reaching pre-pandemic levels.
  • The resurgence is partly attributed to the emergence of “AI PCs,” which feature onboard AI processing capabilities, with projections suggesting these will represent almost 60% of all PC sales by 2027.
  • Major PC manufacturers like Lenovo, HP, Dell, and Apple are heavily investing in the AI PC segment, with Lenovo leading the market and Apple experiencing the fastest growth in shipments.
  • Source

🤖Stability AI launches multilingual Stable LM 2 12B

Stability AI has released a 12-billion-parameter version of its Stable LM 2 language model, offering both a base and an instruction-tuned variant. These models are trained on a massive 2 trillion token dataset spanning seven languages: English, Spanish, German, and more. Stability AI has also improved its 1.6 billion-parameter Stable LM 2 model with better conversational abilities and tool integration.

The new 12B model is designed to balance high performance with relatively lower hardware requirements than other large language models. Stability AI claims it can handle complex tasks requiring substantially more computational resources. The company also plans to release a long-context variant of these models on the Hugging Face platform soon.

Why does this matter?

Stable LM 2 uses powerful 12B models without the most advanced hardware, making it a great choice for enterprises and developers. Stability AI’s multi-pronged approach to language solutions may give it an edge in the competitive generative AI market.

Source

📱 Ferret-UI beats GPT-4V in mobile UI tasks

Researchers have launched Ferret-UI, a multimodal language model designed to excel at understanding and interacting with mobile user interfaces (UIs). Unlike general-purpose models, Ferret-UI is trained explicitly for various UI-centric tasks, from identifying interface elements to reasoning about an app’s overall functionality.

Ferret-UI beats GPT-4V in mobile UI tasks
Ferret-UI beats GPT-4V in mobile UI tasks

By using “any resolution” technology and a meticulously curated dataset, Ferret-UI digs deep into the intricacies of mobile UI screens, outperforming its competitors in elementary and advanced tasks. Its ability to execute open-ended instructions may make it the go-to solution for developers looking to create more intuitive mobile experiences.

Why does this matter?

Ferret-UI’s advanced capabilities in understanding and navigating mobile UI screens will increase accessibility, productivity, and user satisfaction. By setting a new standard for mobile UI interaction, this innovative MLLM paves the way for more intuitive and responsive mobile experiences for users to achieve more with less effort.

Source

⏰ Musk says AI will outsmart humans within a year

Tesla CEO Elon Musk has boldly predicted that AI will surpass human intelligence as early as next year or by 2026. In a wide-ranging interview, Musk discussed AI development’s challenges, including chip shortages and electricity supply constraints, while sharing updates on his xAI startup’s AI chatbot, Grok. Despite the hurdles, Musk remains optimistic about the future of AI and its potential impact on society.

Why does this matter?

Musk’s prediction highlights the rapid pace of AI development and its potential to reshape our world in the near future. As AI becomes increasingly sophisticated, it could transform the job market and raise important ethical questions about the role of technology in society.

Source

What Else Is Happening in April 09th 2024❗

🇬🇧 Microsoft is opening a new AI research hub in London

Microsoft is tapping into the UK’s exceptional talent pool to drive language models and AI infrastructure breakthroughs. The move highlights Microsoft’s commitment to invest £2.5 billion in upskilling the British workforce and building the AI-driven future. (Link)

🎥 OpenAI is using YouTube for GPT-4 training

OpenAI reportedly transcribed over a million hours of YouTube videos to train its advanced GPT-4 language model. Despite legal concerns, OpenAI believes this is fair use. Google and Meta have also explored various solutions to obtain more training data, including using copyrighted material and consumer data. (Link)

🧠 Arm’s new chips bring AI to the IoT edge

Arm has introduced the Ethos-U85 NPU and Corstone-320 IoT platform, designed to enhance edge AI applications with improved performance and efficiency. These technologies aim to accelerate the development and deployment of intelligent IoT devices by providing an integrated hardware and software solution for Arm’s partners. (Link)

🍁 Canada bets big on AI with $2.4B investment

Prime Minister Justin Trudeau has announced a $2.4 billion investment in Canada’s AI sector, with the majority aimed at providing researchers access to computing capabilities and infrastructure. The government also plans to establish an AI Safety Institute and an Office of the AI and Data Commissioner to ensure responsible development and regulation of the technology. (Link)

A Daily chronicle of AI Innovations April 08th 2024: 🇬🇧 Microsoft opens AI Hub in London to ‘advance state-of-the-art language models’ 💡 JPMorgan CEO compares AI’s potential impact to electricity and the steam engine 🎵 Spotify moves into AI with new feature ⚖️ Build resource-efficient LLMs with Google’s MoD 📡 Newton brings sensor-driven intelligence to AI models 💰 Internet archives become AI training goldmines for Big Tech

Build resource-efficient LLMs with Google’s MoD

Google DeepMind has introduced “Mixture-of-Depths” (MoD), an innovative method that significantly improves the efficiency of transformer-based language models. Unlike traditional transformers that allocate the same amount of computation to each input token, MoD employs a “router” mechanism within each block to assign importance weights to tokens. This allows the model to strategically allocate computational resources, focusing on high-priority tokens while minimally processing or skipping less important ones.

Build resource-efficient LLMs with Google's MoD
Build resource-efficient LLMs with Google’s MoD

Notably, MoD can be integrated with Mixture-of-Experts (MoE), creating a powerful combination called Mixture-of-Depths-and-Experts (MoDE). Experiments have shown that MoD transformers can maintain competitive performance while reducing computational costs by up to 50% and achieving significant speedups during inference.

Why does this matter?

MoD can greatly reduce training times and enhance model performance by dynamically optimizing computational resources. Moreover, it adapts the model’s depth based on the complexity of the task at hand. For simpler tasks, it employs shallower layers, conserving resources. Conversely, for intricate tasks, it deepens the network, enhancing representation capacity. This adaptability ensures that creators can fine-tune LLMs for specific use cases without unnecessary complexity.

Source

Newton brings sensor-driven intelligence to AI models

Startup Archetype AI has launched with the ambitious goal of making the physical world understandable to artificial intelligence. By processing data from a wide variety of sensors, Archetype’s foundational AI model called Newton aims to act as a translation layer between humans and the complex data generated by the physical world.

Using plain language, Newton will allow people to ask questions and get insights about what’s happening in a building, factory, vehicle, or even the human body based on real-time sensor data. The company has already begun pilot projects with Amazon, Volkswagen, and healthcare researchers to optimize logistics, enable smart vehicle features, and track post-surgical recovery. Archetype’s leadership team brings deep expertise from Google’s Advanced Technology and Products (ATAP) division.

Why does this matter?

General-purpose AI systems like Newton that can interpret diverse sensor data will be the pathway to building more capable, context-aware machines. In the future, users may increasingly interact with AI not just through screens and speakers but through intelligently responsive environments that anticipate and adapt to their needs. However, as AI becomes more deeply embedded in the physical world, the stakes of system failures or unintended consequences become higher.

Source

Internet archives become AI training goldmines for Big Tech

To gain an edge in the heated AI arms race, tech giants Google, Meta, Microsoft, and OpenAI are spending billions to acquire massive datasets for training their AI models. They are turning to veteran internet companies like Photobucket, Shutterstock, and Freepik, who have amassed vast archives of images, videos, and text over decades online.

The prices for this data vary depending on the type and buyer but range from 5 cents to $7 per image, over $1 per video, and around $0.001 per word for text. The demand is so high that some companies are requesting billions of videos, and Photobucket says it can’t keep up.

Why does this matter?

This billion-dollar rush for AI training data could further solidify Big Tech’s dominance in artificial intelligence. As these giants hoard the data that’s crucial for building advanced AI models, it may become increasingly difficult for startups or academic labs to compete on a level playing field. We need measures to protect the future diversity and accessibility of AI technologies.

Source

🎵 Spotify moves into AI with new feature

  • Spotify is launching a beta tool enabling Premium subscribers to create playlists using text descriptions on mobile.
  • Users can input various prompts reflecting genres, moods, activities, or even movie characters to receive a 30-song playlist tailored to their request, with options for further refinement through additional prompts.
  • The AI Playlist feature introduces a novel approach to playlist curation, offering an efficient and enjoyable way to discover music that matches specific aesthetics or themes, despite limitations on non-music related prompts and content restrictions.
  • Source

🇬🇧 Microsoft opens AI Hub in London to ‘advance state-of-the-art language models’

  • Mustafa Suleyman, co-founder of DeepMind and new CEO of Microsoft AI, announced the opening of a new AI hub in London, focusing on advanced language models, under the leadership of Jordan Hoffmann.
  • The hub aims to recruit fresh AI talent for developing new language models and infrastructure, bolstered by Microsoft’s £2.5 billion investment in the U.K. over the next three years to support AI economy training and data centre expansion.
  • Suleyman, Hoffmann, and about 60 AI experts recently joined Microsoft through its indirect acquisition of UK-based AI startup Inflection AI.
  • Source

💡 JPMorgan CEO compares AI’s potential impact to electricity and the steam engine

  • JPMorgan CEO Jamie Dimon stated AI could significantly impact every job, comparing its potential to revolutionary technologies like the steam engine and electricity.
  • Dimon highlighted AI’s importance in his shareholder letter, revealing the bank’s investment in over 400 AI use cases and the acquisition of thousands of AI experts and data scientists.
  • He expressed belief in AI’s transformative power, equating its future impact to historical milestones such as the printing press, computing, and the internet.
  • Source

What Else Is Happening in AI on April 08th, 2024❗

🎧 Spotify introduces AI-generated personalized playlists

Spotify has launched AI-powered personalized playlists that users can create using text prompts. The feature is currently available in beta for UK and Australia users on iOS and Android. Spotify uses LLMs to understand the prompt’s intent and its personalization technology to generate a custom playlist, which users can further refine. (Link)

🔍 Meta expands “Made with AI” labeling to more content types

Meta will start applying a “Made with AI” badge to a broader range of AI-generated content, including videos, audio, and images. The company will label content where it detects AI image indicators or when users acknowledge uploading AI-generated content. (Link)

🚀 Gretel’s Text-to-SQL dataset sets new standard for AI training data

Gretel has released the world’s largest open-source Text-to-SQL dataset containing over 100,000 high-quality synthetic samples spanning 100 verticals. The dataset, generated using Gretel Navigator, aims to help businesses unlock the potential of their data by enabling AI models to understand natural language queries and generate SQL queries. (Link)

💾 Microsoft upgrades Azure AI Search with more storage and support for OpenAI apps

Microsoft has made Azure AI Search more cost-effective for developers by increasing its vector and storage capacity. The service now supports OpenAI applications, including ChatGPT and GPTs, through Microsoft’s retrieval augmented generation system. Developers can now scale their apps to a multi-billion vector index within a single search without compromising speed or performance. (Link)

📱 Google brings Gemini AI chatbot to Android app

Google is bringing its AI chatbot, Gemini, to the Android version of the Google app. Similar to its iOS integration, users can access Gemini by tapping its logo at the top of the app, opening a chatbot prompt field. Here, users can type queries, request image generation, or ask for image analysis. (Link)

A Daily chronicle of AI Innovations April 06th 2024: 👀 Sam Altman and Jony Ive seek $1B for personal AI device 🚕 Elon Musk says Tesla will unveil robotaxi in August 🔖 Meta to label content ‘made with AI’ 🙃 How OpenAI, Google and Meta ignored corporate policies to train their AI 🛒

👀 Sam Altman and Jony Ive seek $1B for personal AI device OpenAI CEO

Sam Altman and former Apple design chief Jony Ive are collaborating to create an AI-powered personal device and are currently seeking funding. The specifics of the device are unclear, but it is noted to not resemble a smartphone, with speculation about it being similar to the screenless Humane AI pin. The venture, still unnamed, aims to raise up to $1 billion and is in discussions with major investors, including Thrive Capital and Emerson Collective, with potential ownership involvement from OpenAI. https://invest.radintel.ai

🚕 Elon Musk says Tesla will unveil robotaxi in August

Elon Musk announced that Tesla will unveil its robotaxi on August 8th, aiming to focus on autonomous vehicles over mass-market EVs. The Tesla robotaxi is part of Musk’s vision for a shared fleet that owners can monetize, described in the Tesla Network within his Master Plan Part Deux. Musk’s history of ambitious claims about self-driving technology contrasts with regulatory scrutiny and safety concerns involving Tesla’s Autopilot and Full Self-Driving features.

OpenAI’s AI model can clone your voice in 15 seconds

OpenAI has offered a glimpse into its latest breakthrough – Voice Engine, an AI model that can generate stunningly lifelike voice clones from a mere 15-second audio sample and a text input. This technology can replicate the original speaker’s voice, opening up possibilities for improving educational materials.

Though the model has many applications, the AI giant is cautious about its potential misuse, especially during elections. They have strict rules for partners, like no unauthorized impersonation, clear labeling of synthetic voices, and technical measures like watermarking and monitoring.

Meta to label content ‘made with AI’

  • Meta announced that starting in May 2024, AI-generated content on Facebook, Instagram, and Threads will be labeled “Made with AI.”
  • The decision for broader labeling, including AI-generated videos, audio, and images, is influenced by expert consultations and public opinion surveys.
  • Meta’s goal with the “Made with AI” label is to provide more context to users, aiding in content evaluation, while content violating community standards will still be removed.
  • Source

How OpenAI, Google and Meta ignored corporate policies to train their AI

  • OpenAI, Google, and Meta pushed the boundaries of data acquisition for AI development, with OpenAI transcribing over one million hours of YouTube videos for its GPT-4 model.
  • Meta considered extreme measures such as purchasing a publishing house for access to copyrighted materials, and Google amended its privacy policy to potentially harness user-generated content in Google Docs for AI.
  • As the demand for data outpaces supply, tech companies are exploring the creation of synthetic data generated by AI models themselves, despite the risk of models reinforcing their own errors, suggesting a future where AI might train on data it generates.
  • Source

🛒 Tech giants are on a billion-dollar shopping spree for AI training data

  • Tech giants are spending billions to license images, videos, and other content from companies such as Photobucket and Shutterstock to train their AI models, with costs ranging from 5 cents to $1 per photo and more for videos.
  • Prices for licensing data to train AI vary, with figures from $1 to $2 per image, $2 to $4 for short videos, and up to $300 per hour for longer films, while special handling items like nude images may cost $5 to $7 each.
  • Legal concerns arise as companies like Photobucket update their terms of service to sell user-uploaded content for AI training, despite the US Federal Trade Commission warning against retroactively changing terms for AI use, leading to investigations into deals like Reddit’s with Google.
  • Source

A daily chronicle of AI Innovations April 05th 2024: 🤷‍♀️ YouTube CEO warns OpenAI that training models on its videos is against the rules; 🏢 OpenAI says 2024 is the “year of the enterprise” when it comes to AI; ⚔️ The war for AI talent has begun; 🏢 Cohere launches the “most powerful LLM for enterprises”; 🧰 OpenAI doubles down on AI model customization; 🏠 Will personal home robots be Apple’s next big thing?

Cohere launches the “most powerful LLM for enterprises”

Cohere has announced the release of Command R+, its most powerful and scalable LLM to date. Designed specifically for enterprise use cases, Command R+ boasts several key features:

  • Advanced Retrieval Augmented Generation (RAG) to access and process vast amounts of information, improving response accuracy and reliability.
  • Support for ten business languages, enabling seamless operation across global organizations.
  • Tool Use feature to automate complex workflows by interacting with various software tools.

Moreover, Command R+ outperforms other scalable models on key metrics while providing strong accuracy at lower costs.

Cohere launches the “most powerful LLM for enterprises”
Cohere launches the “most powerful LLM for enterprises”

The LLM is now available through Cohere’s API and can be deployed on various cloud platforms, including Microsoft Azure and Oracle Cloud Infrastructure.

Why does this matter?

As one of the first “enterprise-hardened” LLMs optimized for real-world use cases, Command R+ could shape how companies operationalize generative AI across their global operations and product lines. Similar to how Robotic Process Automation (RPA) transformed back-office tasks, Command R+ could significantly improve efficiency and productivity across diverse industries. Additionally, availability on Microsoft Azure and upcoming cloud deployments make it readily accessible to businesses already using these platforms, which could lower the barrier to entry for implementing gen AI solutions.

Source

OpenAI doubles down on AI model customization

OpenAI is making significant strides in AI accessibility with new features for its fine-tuning API and an expanded Custom Models program. These advancements give developers greater control and flexibility when tailoring LLMs for specific needs.

The fine-tuning AP includes:

  • Epoch-based checkpoint creation for easier retraining
  • A playground for comparing model outputs
  • Support for third-party integration
  • Hyperparameters adjustment directly from the dashboard

The Custom Models program now offers assisted fine-tuning with OpenAI researchers for complex tasks and custom-trained models built entirely from scratch for specific domains with massive datasets.

Why does this matter?

This signifies a significant step towards more accessible and powerful AI customization. Previously, fine-tuning required technical expertise and large datasets. Now, with OpenAI’s assisted programs, organizations can achieve similar results without needing in-house AI specialists, potentially democratizing access to advanced AI capabilities.

Source

Will personal home robots be Apple’s next big thing?

Apple is reportedly venturing into personal robotics after abandoning its self-driving car project and launching its mixed-reality headset. According to Bloomberg’s sources, the company is in the early stages of developing robots for the home environment.

Two potential robot designs are mentioned in the report. One is a mobile robot that can follow users around the house. The other is a stationary robot with a screen that can move to mimic a person’s head movements during video calls. Apple is also considering robots for household tasks in the long term.

The project is being spearheaded by Apple’s hardware and AI teams under John Giannandrea. Job postings on Apple’s website further support its commitment to robotics, highlighting its search for talent to develop “the next generation of Apple products” powered by AI.

Why does this matter?

If Apple does release personal home robots, it could mainstream consumer adoption and create new use cases, as the iPhone did for mobile apps and smart assistants. Apple’s brand power and integrated ecosystem could help tackle key barriers like cost and interoperability that have hindered household robotics so far.

It could also transform homes with mobile AI assistants for tasks like elderly care, household chores, entertainment, and more. This may spur other tech giants to double down on consumer robotics.

Source

🤷‍♀️ YouTube CEO warns OpenAI that training models on its videos is against the rules

  • YouTube CEO Neal Mohan warned that OpenAI’s use of YouTube videos to train its text-to-video generator Sora could breach the platform’s terms of service, emphasizing creators’ expectations of content use compliance.
  • This stance poses potential challenges for Google, facing multiple lawsuits over alleged unauthorized use of various content types to train its AI models, arguing such use constitutes “fair use” through transformative learning.
  • Mohan’s remarks could undermine Google’s defense in ongoing legal battles by highlighting inconsistencies in the company’s approach to using content for AI training, including its use of YouTube videos and content from other platforms.
  • Source

⚔️ The war for AI talent has begun

  • Elon Musk aims to retain Tesla’s AI talent by increasing their compensation to counteract aggressive recruitment tactics from OpenAI.
  • Tesla Staff Machine Learning Scientist Ethan Knight’s move to Musk’s AI startup, xAI, exemplifies efforts to prevent employees from joining competitors like OpenAI.
  • Musk describes the ongoing competition for AI professionals as the “craziest talent war” he has ever seen and sees increased compensation as a means to achieve Tesla’s ambitious AI goals, including autonomous driving and humanoid robots development.
  • Source

🏢 OpenAI says 2024 is the “year of the enterprise” when it comes to AI

  • OpenAI’s ChatGPT Enterprise has attracted over 600,000 sign-ups, prompting COO Brad Lightcap to declare 2024 as the “year of adoption for AI in the enterprise”.
  • Despite the strong uptake of ChatGPT Enterprise, OpenAI faces stiff competition from companies eager to penetrate the workplace AI market, including major investor Microsoft with its enterprise AI solutions.
  • OpenAI’s venture into the enterprise sector, especially with ChatGPT Enterprise, marks a significant move towards profitability, with successful partnerships with major media companies like Axel Springer SE, Le Monde, and Prisa.
  • Source

What Else Is Happening in AI on April 05th, 2024❗

📈S&P Global launches AI benchmarking tool

S&P Global has launched S&P AI Benchmarks by Kensho, a groundbreaking tool that evaluates the performance of LLMs in complex financial and quantitative applications. This solution aims to set a new industry standard and promote transparency in AI adoption within the financial sector. (Link)

🤝Waymo and Uber partner for autonomous food delivery in Phoenix

Waymo and Uber have teamed up to launch autonomous Uber Eats deliveries in Phoenix using Waymo’s self-driving vehicles. The service will initially cover select merchants in Chandler, Tempe, and Mesa. Customers can opt out during checkout if they prefer a human courier and will receive instructions for retrieving their order from the autonomous vehicle upon arrival. (Link)

🔍Storyblocks integrates AI for smarter search

Storyblocks has integrated OpenAI’s LLM into its search engine to improve search accuracy for complex queries. Coupled with algorithms analyzing content performance and user engagement, the AI-driven search adapts to provide fresh, high-quality content. Storyblocks also uses machine learning to optimize thumbnails, prioritize representation, and suggest complementary assets, streamlining the creative process. (Link)

🚀Hercules AI streamlines enterprise AI app development

Hercules AI has introduced a new “assembly line” approach for rapid deployment of AI assistants in enterprises. The pre-configured components allow companies to develop cost-effective, scalable AI agents. Plus, their RosettaStoneLLM, built on Mistral-7B and WizardCoder-13B, outperforms competitors by converting data for internal AI workflows. (Link)

🤖Yum Brands embraces AI across restaurants

Yum Brands, the parent company of KFC, Pizza Hut, and Taco Bell, is infusing AI into every aspect of its restaurant operations. From voice AI taking drive-thru orders to an AI-powered “SuperApp” for staff, Yum aims to elevate customer experiences and streamline processes. The AI-driven initiatives include personalized promotions, predictive ordering, and even AI-assisted cooking instructions. (Link)

A daily chronicle of AI Innovations April 04th 2024: 🎵 What’s new in Stability AI’s Stable Audio 2.0? 🖥️ Opera One browser becomes the first to offer local AI integration 🚀 Copilot gets GPT-4 Turbo upgrade
🤖 SWE-agent: AI coder that solves GitHub issues in 93 seconds
📲 Mobile-first Higgsfield aims to disrupt video marketing with AI

What’s new in Stability AI’s Stable Audio 2.0?

Stability AI has released Stable Audio 2.0, a new AI model that generates high-quality, full-length audio tracks. Built upon its predecessor, the latest model introduces three groundbreaking features:

  • Generates tracks up to 3 minutes long with coherent musical structure
  • Enables audio-to-audio generation, allowing users to transform uploaded samples using natural language prompts
  • Enhances sound effect generation and style transfer capabilities, offering more flexibility and control for artists

Stable Audio 2.0’s architecture combines a highly compressed autoencoder and a diffusion transformer (DiT) to generate full tracks with coherent structures. The autoencoder condenses raw audio waveforms into shorter representations, capturing essential features, while the DiT excels at manipulating data over long sequences. This combination allows the model to recognize and reproduce the large-scale structures essential for creating high-quality musical compositions.

Trained exclusively on a licensed dataset from AudioSparx, Stable Audio 2.0 prioritizes creator rights by honoring opt-out requests and ensuring fair compensation. You can explore the capabilities of the model for free on the Stable Audio website.

Why does this matter?

Stable Audio 2’s capability to generate 3-minute songs is a big step forward for AI music tools. But it still has some issues, like occasional glitches and “soulless” vocals, showing that AI has limits in capturing the emotion of human-made music. Also, a recent open letter from artists like Billie Eilish and Katy Perry raises concerns about the ethics of AI-generated music.

Source

SWE-agent: AI coder that solves GitHub issues in 93 seconds

Researchers at Princeton University have developed SWE-agent, an AI system that converts language models like GPT-4 into autonomous software engineering agents. SWE-agent can identify and fix bugs and issues in real-world GitHub repositories in 93 seconds! It does so by interacting with a specialized terminal, which allows it to open, scroll, and search through files, edit specific lines with automatic syntax checking, and write and execute tests. This custom-built agent-computer interface is critical for the system’s strong performance.

SWE-agent: AI coder that solves GitHub issues in 93 seconds
SWE-agent: AI coder that solves GitHub issues in 93 seconds

In the SWE-Bench benchmark test, SWE-agent solved 12.29% of the problems presented, nearly matching the 13.86% achieved by Devin, a closed-source $21 million commercial AI programmer developed by Cognition AI. While Devin is currently only available to select developers, the Princeton team has made SWE-agent open-source to gather feedback and encourage collaboration in advancing this technology.

Why does this matter?

The rise of SWE-agent shows AI systems are becoming more sophisticated in assisting human programmers. Over time, they may change the nature of software development roles, requiring developers to focus more on high-level problem-solving and architectural design while delegating routine tasks to AI assistants. This change could make software development faster and more creative, but it might also require significant upskilling within the developer community.

Source

Mobile-first Higgsfield aims to disrupt video marketing with AI

Former Snap AI chief Alex Mashrabov has launched a new startup called Higgsfield AI, which aims to make AI-powered video creation accessible to creators and marketers. The company’s first app, Diffuse, allows users to generate original video clips from text descriptions or edit existing videos to insert themselves into the scenes.

Higgsfield is taking on OpenAI’s Sora video generator but targeting a broader audience with its mobile-first, user-friendly tools. The startup has raised $8 million in seed funding and plans to further develop its video editing capabilities and AI models. While questions remain around data usage and potential for abuse, Higgsfield believes it can carve out a niche in social media marketing with its realistic, easy-to-use video generation.

Why does this matter?

Higgsfield’s mobile-first approach to AI video generation could be a game-changer regarding accessibility and ease of use. The company is positioning itself to capture a significant portion of the creator economy by prioritizing consumer-friendly features and social media integration. As more users embrace these tools, we can expect to see an explosion of AI-generated content across social media platforms, which could have far-reaching implications for content authenticity and user engagement.

Source

Generative AI Used To Develop Potential New Drugs For Antibiotic-Resistant Bacteria

Researchers at Stanford Medicine and McMaster University have devised a new AI model, SyntheMol (“synthesizing molecules”), which creates recipes for chemists to synthesize drugs in the lab. With nearly 5 million deaths linked to antibiotic resistance globally every year, new ways to combat resistant bacterial strains are urgently needed, according to the researchers.

Using SyntheMol, the researchers have so far developed six novel drugs aimed at killing resistant strains of Acinetobacter baumannii, one of the leading pathogens responsible for antibacterial resistance-related deaths, as noted in a study published March 22 in the journal Nature Machine Intelligence.
Read more here

🤖 Apple explores making personal robots

  • Apple is investigating personal robotics as a new venture, focusing on a mobile robot that can follow users and a robotic table-top device that moves a display around, despite the uncertain future of these products.
  • This move into robotics is part of Apple’s search for new growth avenues after discontinuing its electric vehicle project, with the company looking to capitalize on advancements in artificial intelligence for home automation.
  • Apple’s robotics efforts are led within its hardware engineering division and AI group, indicating a strategic investment in developing cutting-edge home devices, although the projects are still in early research stages and have not been officially confirmed for release.
  • Source

💰 Google could soon start charging a fee for AI-powered search results

  • Google is exploring the introduction of a paid “premium” tier for its search engine, featuring new generative AI-powered enhancements, marking a significant shift from its traditionally ad-supported model.
  • The company is considering integrating these AI-powered search features into existing premium subscription services, amidst concerns about the impact of AI on its advertising revenue, which is critical to its business model.
  • Google has begun experimenting with AI-powered search services, presenting detailed answers alongside traditional search results and advertisements, but has yet to fully implement these features into its main search engine.
  • Source

🖼 ChatGPT now lets you edit AI images created in DALL-E 

  • OpenAI has updated DALL-E with image editing tools accessible within ChatGPT on both web and mobile platforms, allowing users to refine AI-generated images without leaving the chat interface.
  • DALL-E now provides preset style suggestions, such as woodcut, gothic, synthwave, and hand-drawn, to inspire users in their image creation process, similar to AI-generated wallpaper prompts on Android.
  • The integration of DALL-E with ChatGPT, particularly with the latest updates, aims to enhance user-friendliness by simplifying the image creation process and offering starting points for creativity.
  • Source

Meta’s AI image generator struggles to create images of couples of different races. LINK

OpenAI’s Sora just made its first music video and it’s like a psychedelic trip. LINK

What Else Is Happening in AI on April 04th, 2024❗

👨‍💻 Codiumate offers secure, compliant AI-assisted coding for enterprises

Codium AI, an Israeli startup, has launched Codiumate, a semi-autonomous AI agent, to help enterprise software developers with coding, documentation, and testing. It can help with creating development plans from existing code, writing code, finding duplicate code, and suggesting tests. Codiumate aims to make development faster and more secure, with features like zero data retention and the ability to run on private servers or air-gapped computers. (Link)

🖥️ Opera One browser becomes the first to offer local AI integration

Opera now supports 150 local LLM variants in its Opera One browser, making it the first major browser to offer access to local AI models. This feature lets users process their input locally without sending data to a server. Opera One Developer users can select and download their preferred local LLM, which typically requires 2-10 GB of storage space per variant, instead of using Opera’s native browser AI, Aria. (Link)

🧠 AWS expands Amazon Bedrock with Mistral Large model

AWS has included Mistral Large in its Amazon Bedrock managed service for generative AI and app development. Mistral Large is fluent in English, French, Spanish, German, and Italian, and can handle complex multilingual tasks like text understanding, transformation, and code generation. AWS also mentioned that Mistral AI will use its Tranium and Inferentia silicon chips for future models, and that Amazon Bedrock is now in France. (Link)

🚀 Copilot gets GPT-4 Turbo upgrade and enhanced image generation

Microsoft is providing GPT-4 Turbo access to business subscribers of its AI-powered Copilot assistant, without daily limits on chat sessions. The company is also improving image generation capabilities in Microsoft Designer for Copilot subscribers, increasing the limit to 100 images per day using OpenAI’s DALL-E 3 model. These upgrades are part of the $30 per user, per month pricing of Copilot for Microsoft 365. (Link)

🌐 Status invests in Matrix to create a decentralized messaging platform

Status, a mobile Ethereum client, has invested $5 million in New Vector, the company behind the open-source, decentralized communication platform Matrix.org. They plan to create a secure messaging solution for users to control their data and communicate across apps and networks. (Link)

A daily chronicle of AI Innovations April 03rd 2024: 🔍 Google’s Gecko: LLM-powered text embedding breakthrough; 🔓 Anthropic’s “many-shot jailbreaking” wears down AI ethics; 🌌 CosmicMan enables the photorealistic generation of human images

Google’s Gecko: LLM-powered text embedding breakthrough

Gecko is a compact and highly versatile text embedding model that achieves impressive performance by leveraging the knowledge of LLMs. DeepMind researchers behind Gecko have developed a novel two-step distillation process to create a high-quality dataset called FRet using LLMs. The first step involves using an LLM to generate diverse, synthetic queries and tasks from a large web corpus. In the second step, the LLM mines positive and hard negative passages for each query, ensuring the dataset’s quality.

Google's Gecko: LLM-powered text embedding breakthrough
Google’s Gecko: LLM-powered text embedding breakthrough

When trained on FRet combined with other academic datasets, Gecko outperforms existing models of similar size on the Massive Text Embedding Benchmark (MTEB). Remarkably, the 256-dimensional version of Gecko surpasses all models with 768 dimensions, and the 768-dimensional Gecko competes with models that are 7x larger or use embeddings with 5x higher dimensions.

Why does it matter?

Text embedding models are crucial in natural language processing tasks such as document retrieval, sentence similarity, and classification. Gecko’s development shows the potential for creating a single model that can support multiple downstream tasks, eliminating the need for separate embedding models for each task. Using LLMs and knowledge distillation techniques, Gecko achieves strong retrieval performance and sets a strong baseline as a zero-shot embedding model.

Source

Anthropic’s “many-shot jailbreaking” wears down AI ethics 

Researchers at Anthropic discovered a new way to get advanced AI language models to bypass their safety restrictions and provide unethical or dangerous information. They call this the “many-shot jailbreaking” technique. By including many made-up dialog examples in the input where an AI assistant provides harmful responses, the researchers could eventually get the real AI to override its training and provide instructions on things like bomb-making.

Anthropic’s “many-shot jailbreaking” wears down AI ethics 
Anthropic’s “many-shot jailbreaking” wears down AI ethics

The researchers say this vulnerability arises from AI models’ increasing ability to process and “learn” from very long input sequences. Essentially, the AI mimics the unethical behavior repeatedly demonstrated in the made-up examples. Anthropic has implemented safeguards against this attack on its systems and has also shared the findings openly so other AI companies can work on mitigations.

Why does it matter?

As AI models become more capable over time, techniques to override their built-in ethical restraints pose serious risks if not addressed. While Anthropic has been transparent in disclosing this vulnerability to enable mitigations, it underscores the need for continued research into AI safety and security. Simple precautions like limiting input length are inadequate; more sophisticated AI “jailbreak” prevention methods are required as these systems advance.

Source

CosmicMan enables the photorealistic generation of human images 

Researchers at the Shanghai AI Laboratory have created a new AI model called CosmicMan that specializes in generating realistic images of people. CosmicMan can produce high-quality, photorealistic human images that precisely match detailed text descriptions, unlike current AI image models that struggle with human images.

CosmicMan enables the photorealistic generation of human images 
CosmicMan enables the photorealistic generation of human images

The key to CosmicMan’s success is a massive dataset called CosmicMan-HQ 1.0 containing 6 million annotated human images and a novel training method—“ Annotate Anyone,” which focuses the model on different parts of the human body. By categorizing words in the text description into body part groups like head, arms, legs, etc., the model can generate each part separately for better accuracy and customizability, thereby outperforming the current state-of-the-art models.

CosmicMan enables the photorealistic generation of human images 
CosmicMan enables the photorealistic generation of human images

Why does it matter?

Existing AI models have struggled to create realistic human images and accurately represent diverse human appearances. With CosmicMan, AI systems will be better equipped to generate high-fidelity images of people, which can have implications for computer vision, graphics, entertainment, virtual reality, and fashion. It may enable more realistic virtual avatars, improved character generation in games and movies, and enhanced visual content creation.

Source

OpenAI-Superhuman introduces a new era of email with OpenAI.

 OpenAI-Superhuman introduces a new era of email with OpenAI.
OpenAI-Superhuman introduces a new era of email with OpenAI

Source

Apple Vision Pro’s Spatial Avatars are a game changer

Get the Meta Quest 3 at half the price for similar functionalities here

Meta Quest 3

UBTECH and Baidu have partnered to integrate large AI models into humanoid robots. Their demo features the Walker S robot folding clothes and sorting objects through natural language, using Baidu’s LLM, ERNIE Bot, for task interpretation/planning.

UBTECH and Baidu have partnered to integrate large AI models into humanoid robots. Their demo features the Walker S robot folding clothes and sorting objects through natural language, using Baidu’s LLM, ERNIE Bot, for task interpretation/planning.
byu/SharpCartographer831 insingularity

YCombinator’s AI boom is still going strong (W24)

With YC’s latest Demo Day (W24), the AI companies are continuing to grow. Six months ago, there were around 139 companies working with AI or ML – that number has climbed to 158, a clear majority of 65% (there are 243 total companies in the batch).

Let’s dive into what’s new, what’s stayed the same, and what we can learn about the state of AI startups.

YCombinator's AI boom is still going strong (W24)
YCombinator’s AI boom is still going strong (W24)

The biggest domains stayed big

Perhaps unsurprisingly, the most popular categories remained unchanged from the last batch. Last time, the top 4 domains were AI Ops, Developer Tools, Healthcare + Biotech, and Finance + Payments. This time, the top 5 were:

  • Developer Tools: Apps, plugins, and SDKs making it easier to write code. Tools for testing automation, website optimization, codebase search, improved Jupyter notebooks, and AI-powered DevOps were all present. There was also a strong contingent of code-generation tools, from coding Copilots to no-code app builders.
  • AI Ops: Tooling and platforms to help companies deploy working AI models. That includes hosting, testing, data management, security, RAG infrastructure, hallucination mitigation, and more. We’ll discuss how the AI Ops sector has continued to mature below.
  • Healthcare + Biotech: While I’ve once again lumped these two categories together, there’s a pretty big split in the types of AI businesses being built. Healthcare companies are building automation tools for the entire healthcare lifecycle: patient booking, reception, diagnosis, treatment, and follow-up. Whereas biotech companies are creating foundation models to enable faster R&D.
  • Sales + Marketing: Early generative AI companies were focused on the sales and marketing benefits of GPT-3: write reasonable sounding copy instantly. Now, we’re seeing more niche use cases for revenue-generating AI: AI-powered CRMs for investors, customer conversation analysis, and AI personal network analysis were among some sales-oriented companies.
  • Finance: Likewise, on the finance side, companies covered compliance, due diligence, deliverable automation, and more. Perhaps one of my favorite descriptions was “a universal API for tax documents.”

The long tail is getting longer

Even though the top categories were quite similar, one new aspect was a wider distribution of industries. Compared with the last batch, there were roughly 35 categories of companies versus 28 (examples of new categories include HR, Recruiting, and Aerospace). That makes sense to me. I’ve been saying for a while now that “AI isn’t a silver bullet” and that you need domain-expertise to capture users and solve new problems.

But it’s also clear that with AI eating the world, we’re also creating new problems. It was interesting to see companies in the batch focused on AI Safety – one company is working on fraud and deepfake detection, while another is building foundation models that are easy to align. I suspect we will continue seeing more companies dealing with the second-order effects of our new AI capabilities.

We’re also seeing more diverse ways of applying AI. In the last batch, a dominant theme was “copilots.” And while those are still present here (as well as “agents”), there are also more companies building “AI-native” products and platforms – software that uses AI in ways beyond a shoehorned sidebar conversation with an AI assistant.

What comes after CustomGPTs?

“AI agents. These will integrate more fully into numerous systems and you would give them the authority to execute things on your behalf. I.e. making reservations for dinner somewhere and then sending you the details or searching and purchasing and sending a gift to someone or planning and executing a vacation reservation including my purchasing travel arrangements, hotel stays, transport to and from, etc. Even something as simple as telling it you are hungry and having and AI agent find something you would like and having it delivered to you. Or it acting on its own to do any number of those because it also sees your schedule, knows you didn’t really eat all day and that it is your mom’s birthday and you forgot to get her anything or to even call…”

How accurate is that statement above?

AI agents are software entities that act autonomously on behalf of their users, making decisions or performing tasks based on predefined criteria, learned preferences, or adaptive learning algorithms. They can range from simple chatbots to sophisticated systems capable of managing complex tasks. The accuracy of the statement reflects a forward-looking perspective on the capabilities of AI agents, envisioning a future where they are deeply integrated into our daily lives, handling tasks from personal to professional spheres with minimal human intervention.

  • 🤖 Autonomy and Integration: The description is accurate in envisioning AI agents that are more fully integrated into various systems. This integration will likely increase as advancements in AI, machine learning, and data analytics continue to evolve. Such agents will understand user preferences, schedules, and even predict needs based on historical data and real-time inputs.
  • 🔍 Executing Tasks on Behalf of Users: The ability of AI agents to perform tasks such as making reservations, purchasing gifts, or arranging travel is not only plausible but is already being realized to a certain extent with existing AI and machine learning technologies. Examples include virtual assistants like Google Assistant, Siri, and Alexa, which can perform a range of tasks from setting reminders to booking appointments.
  • 🎁 Personalization and Prediction: The statement also touches on the AI agents’ capability to act proactively based on the user’s schedule, preferences, or significant dates. This level of personalization and predictive action is a key area of development in AI, aiming to provide more personalized and anticipative user experiences. Implementing this effectively requires sophisticated models of user behavior and preferences, which can be built using machine learning techniques.
  • 🚀 Future Prospects and Ethical Considerations: While the vision of AI agents acting autonomously to manage aspects of our lives is grounded in realistic expectations of technology’s trajectory, it also raises ethical and privacy concerns. Issues such as data security, user consent, and the potential for over-dependence on technology for personal tasks are significant. The development and deployment of such AI agents must consider these aspects to ensure that they serve users’ interests ethically and securely.
  • 📈 Current Limitations and Challenges: It’s important to note that while the statement captures a future potential, current AI technologies have limitations. The complexity of fully understanding human needs, contexts, and the nuances of personal preferences in an ethical manner remains a challenge.

What Else Is Happening in AI on April 03rd, 2024❗

🎮 Microsoft is planning to add an AI chatbot to Xbox

Microsoft is currently testing a new AI-powered chatbot to be added to Xbox to automate customer support tasks. The software giant has tested an “embodied AI character” that animates when responding to Xbox support queries. The virtual representative can handle either text or voice requests. It’s an effort to integrate AI into Xbox platforms and services. (Link)

☁️ CloudFare launches Workers AI to power one-click deployment with Hugging Face

CloudFare has launched Workers AI, which empowers developers to bring their AI applications from Hugging Face to its platform in one click. The serverless GPU-powered interface is generally available to the public. The Cloudflare-Hugging Face integration was announced nearly seven months ago. It makes it easy for models to be deployed onto Workers AI. (Link)

🍺 Machine Learning can predict and enhance complex beer flavor

In a study by Nature Communications, researchers combined chemical analyses, sensory data, and machine learning to create models that accurately predict beer flavor and consumer appreciation from the beer’s chemical composition. They identified compounds that enhance flavor and used this knowledge to improve the taste and popularity of commercial beers. (Link)

📖 Read AI adds AI summaries to meetings, emails, and messages

Read AI is expanding its services from summarizing video meetings to including messages and emails. The platform connects to popular communication platforms like Gmail, Outlook, Slack, Zoom, Microsoft Teams, and Google Meet to deliver daily updates, summaries, and AI-generated takeaways. The goal is to help users save time and improve productivity. (Link)

🤖 Bille Elish, Kety Perry, and 200 other artists protest AI’s devaluation of music

Nicki Minaj, Billie Eilish, Katy Perry and other musicians warn against replacing human singers with AI

In an open letter, over 200 famous musicians, including Billie Eilish and Katy Perry, have expressed their concerns about the negative impact of AI on human creativity. They call for the responsible use of AI and urge AI companies to stop creating music that undermines their work. They believe that unregulated and uncontrolled use of AI can harm songwriters, musicians, and creators. They emphasize the need to protect artists’ rights and fair compensation. (Link)

A daily chronicle of AI Innovations April 02nd 2024: 📲 Apple’s Siri will now understand what’s on your screen; 🤖 OpenAI introduces instant access to ChatGPT; 🚨 Elon Musk says AI might destroy humanity, but it’s worth the risk; 🤖 Sam Altman gives up control of OpenAI Startup Fund; 📰 Yahoo acquires Instagram co-founders’ AI-powered news startup Artifact

🤖 Sam Altman gives up control of OpenAI Startup Fund

  • Sam Altman has relinquished formal control of the OpenAI Startup Fund, which he initially managed, to Ian Hathaway, marking a resolution to the fund’s unique corporate structure.
  • The fund was established in 2021 with Altman temporarily at the helm to avoid potential conflicts had he not returned as CEO after a brief departure; he did not personally invest in or financially benefit from it.
  • Under Hathaway’s management, the fund, starting with $175 million in commitments, has grown to $325 million in assets and has invested in early-stage AI companies across healthcare, law, education, and more, with at least 16 startups backed.
  • Source

🙏 US and UK sign deal to partner on AI research 

  • The US and UK have formed a partnership focused on advancing the safety testing of AI technologies, sharing information and expertise to develop tests for cutting-edge AI models.
  • A Memorandum of Understanding (MOU) has been signed to enhance the regulation and testing of AI, aiming to effectively assess and mitigate the risks associated with AI technology.
  • The partnership involves the exchange of expert personnel between the US and UK AI Safety Institutes, with plans for potential joint testing on publicly available AI models, reinforcing their commitment to addressing AI risks and promoting its safe development globally.
  • Source

📰 Yahoo acquires Instagram co-founders’ AI-powered news startup Artifact

  • Yahoo is acquiring the AI news app Artifact, built by Instagram co-founders, but not its team, aiming to enhance its own news platform with Artifact’s advanced technology and recommendation systems.
  • Artifact’s technology, which focuses on personalizing and recommending content, will be integrated into Yahoo News and potentially other Yahoo platforms, despite the discontinuation of the Artifact app itself.
  • The integration of Artifact’s technology into Yahoo aims to create a personalized content ecosystem, leveraging Yahoo’s vast user base to realize the potential of AI in news curation and recommendation.
  • Source

Apple’s Siri will now understand what’s on your screen

Apple researchers have developed an AI system called ReALM which enables voice assistants like Siri to understand contextual references to on-screen elements. By converting the complex task of reference resolution into a language modeling problem, ReALM outperforms even GPT-4 in understanding ambiguous references and context.

Apple's Siri will now understand what’s on your screen
Apple’s Siri will now understand what’s on your screen

This innovation lies in reconstructing the screen using parsed on-screen entities and their locations to generate a textual representation that captures the visual layout. This approach, combined with fine-tuning language models specifically for reference resolution, allows ReALM to achieve substantial performance gains compared to existing methods.

  • Apple researchers have developed an AI system called ReALM that can understand screen context and ambiguous references, improving interactions with voice assistants.
  • ReALM reconstructs the screen using parsed on-screen entities to generate a textual representation, outperforming GPT-4.
  • Apple is investing in making Siri more conversant and context-aware through this research.
  • However, automated parsing of screens has limitations, especially with complex visual references.
  • Apple is catching up in AI research but faces stiff competition from tech rivals like Google, Microsoft, Amazon, and OpenAI.

Why does this matter?

ReALM’s ability to understand screen context creates possibilities for more intuitive and hands-free interactions with voice assistants. Imagine effortlessly instructing Siri to “open the app at the bottom right corner.” As Apple races to close the AI gap with rivals like Google and Microsoft, ReALM could be a game-changer in making Siri and other Apple products more contextually aware.

Source

OpenAI introduces instant access to ChatGPT

OpenAI now allows users to use ChatGPT without having to create an account. With over 100 million weekly users across 185 countries, it can now be accessed instantly by anyone curious about its capabilities.

While this move makes AI more accessible, other OpenAI products like DALL-E 3 still require an account. The company has also introduced new content safeguards and allows users to opt out of model training, even without an account. Despite growing competition from rivals like Google’s Gemini, ChatGPT remains the most visited AI chatbot site, attracting 1.6 billion visitors in February.

Why does this matter?

By allowing anyone to instantly access ChatGPT, OpenAI is expanding its user base and encouraging more people to explore the potential applications of AI. This move could accelerate the adoption of AI tools across various industries, as users become more comfortable with the technology.

Source

Elon Musk says AI might destroy humanity, but it’s worth the risk

Elon Musk recently shared his thoughts on the potential dangers of AI at the Abundance Summit’s “Great AI Debate” seminar. He estimated a 10-20% chance that AI could pose an existential threat to humanity.

Despite the risks, Musk believes that the benefits of AI outweigh the potential dangers. He emphasized the importance of teaching AI to be truthful and curious, although he didn’t provide specifics on how he arrived at his risk assessment.

Why does this matter?

Musk’s comments emphasize the importance of using AI’s advantages while addressing its potential risks. This involves creating transparent, accountable AI systems aligned with human values. While his estimate is concerning, continued research in AI safety and governance is necessary to ensure AI remains beneficial.

Source

Artificial intelligence is taking over drug development

The most striking evidence that artificial intelligence can provide profound scientific breakthroughs came with the unveiling of a program called AlphaFold by Google DeepMind. In 2016 researchers at the company had scored a big success with AlphaGo, an ai system which, having essentially taught itself the rules of Go, went on to beat the most highly rated human players of the game, sometimes by using tactics no one had ever foreseen. This emboldened the company to build a system that would work out a far more complex set of rules: those through which the sequence of amino acids which defines a particular protein leads to the shape that sequence folds into when that protein is actually made. AlphaFold found those rules and applied them with astonishing success.

The achievement was both remarkable and useful. Remarkable because a lot of clever humans had been trying hard to create computer models of the processes which fold chains of amino acids into proteins for decades. AlphaFold bested their best efforts almost as thoroughly as the system that inspired it trounces human Go players. Useful because the shape of a protein is of immense practical importance: it determines what the protein does and what other molecules can do to it. All the basic processes of life depend on what specific proteins do. Finding molecules that do desirable things to proteins (sometimes blocking their action, sometimes encouraging it) is the aim of the vast majority of the world’s drug development programmes.

Source

Comment: Someone needs to fire up a CRISPR-cas AI service you can submit your DNA to and they develop and ship you a treatment kit for various cancers, genetic disorders etc.

What Else Is Happening in AI on April 02nd, 2024❗

🚫 Pinecone launches Luna AI that never hallucinates

Trained using a novel “information-free” approach, Luna achieved zero hallucinations by always admitting when it doesn’t know an answer. The catch? Its performance on other tasks is significantly reduced. While not yet open-sourced, vetted institutions can access the model’s source and weights. (Link)

🤝 US and UK  collaborate to tackle AI safety risks

As concerns grow over the potential risks of next-gen AI, the two nations will work together to develop advanced testing methods and share key information on AI capabilities and risks. The partnership will address national security concerns and broader societal issues, with plans for joint testing exercises and personnel exchanges between their respective AI safety institutes. (Link)

🔍 Perplexity to test sponsored questions in AI search

Perplexity’s Chief Business Officer, Dmitry Shevelenko, announced the company’s plan to introduce sponsored suggested questions later this year. When users search for more information on a topic, the platform will display sponsored queries from brands, allowing Perplexity to monetize its AI search platform. (Link)

🇯🇵 OpenAI expands to Japan with Tokyo office

The Tokyo office will be OpenAI’s first in Asia and third international location, following London and Dublin. The move aims to offer customized AI services in Japanese to businesses and contribute to the development of an AI governance framework in the country. (Link)

🤖 Bixby gets a GenAI upgrade

Despite speculation, Samsung isn’t giving up on its voice assistant, Bixby. Instead, the company is working hard to equip Bixby with generative AI to make it smarter and more conversational. Samsung introduced a suite of AI features called Galaxy AI to its smartphones, including the Galaxy S24’s use of Google’s Gemini Nano AI model. (Link)

A daily chronicle of AI Innovations April 01st 2024: 🎤 This AI model can clone your voice in 15 seconds; 🚀 Microsoft and OpenAI plan $100B supercomputer for AI development; 🖼️ MagicLens: Google DeepMind’s breakthrough in image retrieval technology

🍎Apple says its latest AI model is even better than OpenAI’s GPT4

  • Apple researchers have introduced ReALM, an advanced AI model designed to understand and navigate various contexts more effectively than OpenAI’s GPT4.
  • ReALM aims to enhance user interaction by accurately understanding onscreen, conversational, and background entities, making device interactions more intuitive.
  • Apple believes ReALM’s ability to handle complex reference resolutions, including onscreen elements, positions it as a superior solution compared to the capabilities of GPT-4.
 

Deepmind chief doesn’t see AI reaching its limits anytime soon

  • Deepmind founder Demis Hassabis believes AI is both overhyped and underestimated, with the potential for AI far from being reached and warning against the excessive hype surrounding it.
  • Hassabis predicts many AI startups will fail due to the high computing power demands, expects industry consolidation, and sees no limit to the advancements in massive AI models.
  • Despite concerns over hype, Hassabis envisions the beginning of a new golden era in scientific discovery powered by AI and estimates a 50% chance of achieving artificial general intelligence within the next ten years.

This AI model can clone your voice in 15 seconds

OpenAI has offered a glimpse into its latest breakthrough – Voice Engine, an AI model that can generate stunningly lifelike voice clones from a mere 15-second audio sample and a text input. This technology can replicate the original speaker’s voice, opening up possibilities for improving educational materials, making videos more accessible to global audiences, assisting with communication for people with speech impairments, and more.

Reference audio:

LISTEN NOW · 0:15

Generated audio:

LISTEN NOW · 0:16

Though the model has many applications, the AI giant is cautious about its potential misuse, especially during elections. They have strict rules for partners, like no unauthorized impersonation, clear labeling of synthetic voices, and technical measures like watermarking and monitoring. OpenAI hopes this early look will start a conversation about how to address potential issues by educating the public and developing better ways to trace the origin of audio content.

Why does this matter?

OpenAI’s Voice Engine can transform industries from gaming and entertainment to education and healthcare. Imagine video games with non-player characters that sound like real people, animated films with AI-generated voiceovers, or personalized voice assistants for individuals with speech impairments. But as AI-generated voices become more human-like, questions about consent, privacy, and robust authentication measures must be addressed to prevent misuse.

Source

Microsoft+OpenAI plan $100B supercomputer for AI development

Microsoft and OpenAI are reportedly planning to build a massive $100 billion supercomputer called “Stargate” to rapidly advance the development of OpenAI’s AI models. Insiders say the project, set to launch in 2028 and expand by 2030, would be one of the largest investments in computing history, requiring several gigawatts of power – equivalent to multiple large data centers.

Much of Stargate’s cost would go towards procuring millions of specialized AI chips, with funding primarily from Microsoft. A smaller $10B precursor called “Phase 4” is planned for 2026. The decision to move forward with Stargate relies on OpenAI achieving significant improvements in AI capabilities and potential “superintelligence.” If realized, Stargate could enable OpenAI’s AI systems to recursively generate synthetic training data and become self-improving.

Why does this matter?

The Stargate project will give OpenAI and Microsoft a massive advantage in creating AI systems that are far more capable than what we have today. This could lead to breakthroughs in areas like scientific discovery, problem-solving, and the automation of complex tasks. But it also raises concerns about the concentration of power in the AI industry. We’ll need new frameworks for governing advanced AI to ensure it benefits everyone, not just a few giants.

Source

MagicLens: Google DeepMind’s breakthrough in image retrieval technology

Google DeepMind has introduced MagicLens, a revolutionary set of image retrieval models that surpass previous state-of-the-art methods in multimodality-to-image, image-to-image, and text-to-image retrieval tasks. Trained on a vast dataset of 36.7 million triplets containing query images, text instructions, and target images, MagicLens achieves outstanding performance while meeting a wide range of search intents expressed through open-ended instructions.

Multimodality-to-Image performance

MagicLens: Google DeepMind's breakthrough in image retrieval technology
MagicLens: Google DeepMind’s breakthrough in image retrieval technology

Image-to-Image performance

MagicLens employs a dual-encoder architecture, which allows it to process both image and text inputs, delivering highly accurate search results even when queries are expressed in everyday language. By leveraging advanced AI techniques, like contrastive learning and single-modality encoders, MagicLens can satisfy diverse search intents and deliver relevant images with unprecedented efficiency.

Why does this matter?

The release of MagicLens highlights the growing importance of multimodal AI systems that can process both text and visual information. We can expect to see more seamless integration between language and vision, enabling the development of more sophisticated AI applications. This trend could have far-reaching implications for fields such as robotics, autonomous vehicles, and augmented reality, where the ability to interpret and respond to visual data is crucial.

Source

What Else Is Happening in AI on April 01st, 2024❗

🧠 TCS aims to build the largest AI-ready workforce

Tata Consultancy Services (TCS) has announced that it has trained 3.5 lakh employees, more than half of its workforce, in generative AI skills. The company set up a dedicated AI and cloud business unit in 2023 to address the growing needs of customers for cloud and AI adoption, offering a comprehensive portfolio of GenAI services and solutions. (Link)

🔗 ChatGPT introduces hyperlinked source citations in the latest update

OpenAI has introduced a feature for ChatGPT premium users that makes source links more prominent in the bot’s responses. The update hyperlinks words within ChatGPT’s answers, directing users to the source websites — a feature already present in other chatbot search resources like Perplexity. (Link)

✏️ OpenAI’s DALL·E now allows users to edit generated images

OpenAI has launched a new image editing feature for DALL·E, enabling users to modify generated images by selecting areas and describing changes. The editor offers tools to add, remove, or update objects within the image using either the selection tool or conversational prompts. (Link)

🚇 NYC to test Evolv’s AI gun detection technology in subways

New York City plans to test Evolv’s AI-powered gun detection scanners in subway stations within 90 days, according to Mayor Eric Adams. However, Evolv is under scrutiny for the accuracy of its technology, facing reports of false positives and missed detections. (Link)

🚫 Microsoft Copilot banned in US House due to potential data breaches

The US House of Representatives has banned its staffers from using Microsoft Copilot due to concerns about possible data leaks to unauthorized cloud services. This decision mirrors last year’s restriction on the use of ChatGPT in congressional offices, with no other chatbots currently authorized. Microsoft has indicated that it plans to address federal government security and compliance requirements for AI tools like Copilot later this year. (Link)

A Daily Chronicle of AI Innovations in March 2024

  • AI: The Ultimate Sherlocking?
    by /u/mintone (Artificial Intelligence) on July 26, 2024 at 12:16 pm

    submitted by /u/mintone [link] [comments]

  • Speech-to-Text Solution for Multilingual Sentences / Mixed-language speech
    by /u/simbaninja33 (Artificial Intelligence Gateway) on July 26, 2024 at 11:54 am

    I am looking for a speech-to-text solution, either paid or open-source, that can accurately transcribe speech containing a mix of two languages within the same sentence. I have explored options like Microsoft Azure, Google Cloud, and OpenAI, but haven't found a satisfactory solution yet. For example, I need the solution to handle sentences like: "I have tried the restaurant yesterday, it is muy muy bueno, they serve some of the pizza, que haria mi abuela super celoza de la receta." "I went to the store y compré un poco de pan because we were running low." I have already tried Microsoft Azure, which can handle multiple languages, but only when they are not mixed within the same sentence (as mentioned in their documentation). Google Cloud's speech-to-text fails to accurately transcribe mixed-language speech, and OpenAI doesn't seem to offer this functionality. I am open to both continuous real-time speech recognition and file-based recognition. For real-time applications, I am also willing to consider workarounds, such as implementing a "button" that can be clicked to quickly switch between the main language and the second language. If anyone has experience with a solution that can handle this type of mixed-language speech recognition, I would greatly appreciate any suggestions or recommendations. Thank you in advance for your help! submitted by /u/simbaninja33 [link] [comments]

  • Any open source AI model with web search abilities?
    by /u/david8840 (Artificial Intelligence Gateway) on July 26, 2024 at 11:45 am

    Is there any open source AI model with web search abilities? I want to be able to ask it questions which require real time internet searching, for example "What is the weather like now in NY?" submitted by /u/david8840 [link] [comments]

  • Which companies are leading the way in AI detection? (for audio/video deepfakes, etc.?)
    by /u/ProfessionalHat3555 (Artificial Intelligence Gateway) on July 26, 2024 at 11:21 am

    So I was listening to the most recent Bill Simmons pod w/ Derek Thompson where they discuss conspiracy theories and AI shit-detection (40:00-48:00 if you're curious)... 1ST Q: what companies are you aware of that are already working on AI detection? 2ND Q: where do you think the AI detection slice of the market is going? Will there be consumer-grade products that we can use to run, say, a political video through a detection software & get a % of realness rating on it? Will these tools ONLY be available to big conglomerates who become the purveyors of truth? 3RD Q: If we're UNABLE to do this at-scale yet, what would need to happen tech-wise for AI detection to become more accessible to more people? (disclaimer: I'm not a dev) submitted by /u/ProfessionalHat3555 [link] [comments]

  • AI can't take people's jobs if there's no people.
    by /u/baalzimon (Artificial Intelligence Gateway) on July 26, 2024 at 10:53 am

    Looks more and more likely that human populations will decline in the future. Maybe the workforce will just be AI robots rather than young people. PEW: The Experiences of U.S. Adults Who Don’t Have Children 57% of adults under 50 who say they’re unlikely to ever have kids say a major reason is they just don’t want to; 31% of those ages 50 and older without kids cite this as a reason they never had them https://www.pewresearch.org/social-trends/2024/07/25/the-experiences-of-u-s-adults-who-dont-have-children/ submitted by /u/baalzimon [link] [comments]

  • UK School Under Fire for Unlawful Facial-Recognition Use
    by /u/Think_Cat1101 (Artificial Intelligence Gateway) on July 26, 2024 at 10:43 am

    https://www.msn.com/en-us/news/technology/uk-school-under-fire-for-unlawful-facial-recognition-use/ar-BB1qEmeX?cvid=6dfe65854c6e4c2ad473b0e649e795b2&ei=10 submitted by /u/Think_Cat1101 [link] [comments]

  • OpenAI reveals 'SearchGPT'
    by /u/Mindful-AI (Artificial Intelligence Gateway) on July 26, 2024 at 10:41 am

    submitted by /u/Mindful-AI [link] [comments]

  • Amazon’s AI Chip Revolution: How They’re Ditching Nvidia’s High Prices and Speeding Ahead
    by /u/alyis4u (Artificial Intelligence Gateway) on July 26, 2024 at 9:23 am

    Six engineers tested a brand-new, secret server design on a Friday afternoon in Amazon.com’s chip lab in Austin, Texas. Amazon executive Rami Sinno said on Friday during a visit to the lab that the server was full of Amazon’s AI chips, which compete with Nvidia’s chips and are the market leader.https://theaiwired.com/amazons-ai-chip-revolution-how-theyre-ditching-nvidias-high-prices-and-speeding-ahead/ submitted by /u/alyis4u [link] [comments]

  • OpenAI's SearchGPT Is Coming For Google Search; Here Are The Features That Will Reportedly Make It Better
    by /u/vinaylovestotravel (Artificial Intelligence Gateway) on July 26, 2024 at 9:00 am

    Dubbed "SearchGPT," the tool will offer "fast and timely answers with clear and relevant sources" by referencing content from websites and news publishers, including OpenAI content partners such as News Corp (The Post's parent company) and The Atlantic. Read more: https://www.ibtimes.co.uk/openais-searchgpt-coming-google-search-here-are-features-that-will-reportedly-make-it-better-1725770 submitted by /u/vinaylovestotravel [link] [comments]

  • Deleting chats from Blackbox AI?
    by /u/Intelligent-Fig-7791 (Artificial Intelligence Gateway) on July 26, 2024 at 7:40 am

    How on earth do you delete chats from blackbox.ai ? it seems like all chats are public by default submitted by /u/Intelligent-Fig-7791 [link] [comments]

Industries Affected by the Growth of AI Technology

Industries Affected by the Growth of AI Technology

Artificial intelligence (AI) is a pillar technology that’s redefining the way we live, work, and interact. Its influence is particularly pronounced in various industries, bringing about unprecedented changes and challenges. There are some ethical concerns in relation to AI circulating the business world too, though many argue the advantages far outweigh the pitfalls. Here, we’ll explore four industries affected by the growth of AI technology, revealing how it’s dramatically reshaping our professional landscapes and consumer experiences.

The Healthcare Industry

AI’s integration in healthcare significantly enhances diagnostic accuracy, streamlines administrative operations, and personalizes patient care. In diagnostics, AI platforms can sift through vast amounts of medical data to detect anomalies and predict diseases, often surpassing human precision.

Administratively, AI helps optimize workflows, reducing clerical errors and freeing up staff to focus on patient care. Chatbots and virtual assistants are also cropping up and provide round-the-clock support and information to patients. Additionally, AI plays a pivotal role in medical device innovation and drives s manufactures to create revolutionary treatment options.

The Financial Services Sector

The financial sector is also being reinvented by AI. Robo-advisors automate investment advice, while AI algorithms significantly influence the stock market with high-frequency trading and risk management.

Banks and credit card companies are also employing AI to enhance fraud protection and cybersecurity measures. Machine learning models analyze spending patterns and offer personalized banking services, expertly profiling customer’s needs and promoting new products.

Retail and E-commerce

The retail world is also an industry affected by the growth of AI technology as it’s reshaping how retail conducts business, from personalized marketing to optimized supply chain management. Virtual shopping assistants powered by AI help customers find products, while machine learning algorithms predict trends and optimize inventory levels. On the marketing side, AI enables targeted advertising and dynamic pricing strategies.

Retailers also leverage AI to enhance the customer experience through chatbots and recommendation engines. One notable challenge is ensuring that AI systems provide fair and unbiased product recommendations and that they don’t infringe on consumer privacy.

Transportation and Logistics

Transportation and logistics are prime sectors for AI integration as well due to their reliance on complex systems and data-heavy operations. Self-driving vehicles are the most visible AI application but also the most challenging, requiring extensive safety testing and regulatory approval.

AI’s reach extends to route optimization, predictive maintenance, and real-time traffic management, improving efficiency and cutting down on costs. In freight, AI streamlines the loading process and monitors the condition of goods while in transit, enhancing overall supply chain transparency.

AI is a revolutionary technology that we expect to see more of in the future. While there are several ways to benefit from these programs, it’s also vital that industry professionals know how to implement them in balanced ways to minimize challenges. The more you know now, the better you can prepare yourself for integrating AI solutions of your own.

A Daily Chronicle of AI Innovations in March 2024

A Daily Chronicle of AI Innovations in March 2024

AI Innovations in March 2024.

Welcome to the March 2024 edition of the Daily Chronicle, your gateway to the forefront of Artificial Intelligence innovation! Embark on a captivating journey with us as we unveil the most recent advancements, trends, and revolutionary discoveries in the realm of artificial intelligence. Delve into a world where industry giants converge at events like ‘AI Innovations at Work’ and where visionary forecasts shape the future landscape of AI. Stay abreast of daily updates as we navigate through the dynamic realm of AI, unraveling its potential impact and exploring cutting-edge developments throughout this enthralling month. Join us on this exhilarating expedition into the boundless possibilities of AI in March 2024.

Experience the transformative capabilities of AI with “Read Aloud For Me – AI Dashboard” – your ultimate AI Dashboard and Hub. Seamlessly access a comprehensive suite of top-tier AI tools within a single app, meticulously crafted to enhance your efficiency and streamline your digital interactions. Now available on the web at readaloudforme.com and across popular app platforms including Apple, Google, and Microsoft, “Read Aloud For Me – AI Dashboard” places the future of AI at your fingertips, blending convenience with cutting-edge innovation. Whether for professional endeavors, educational pursuits, or personal enrichment, our app serves as your portal to the forefront of AI technologies. Embrace the future today by downloading our app and revolutionize your engagement with AI tools.

Unlock the power of AI with "Read Aloud For Me" – your ultimate AI Dashboard and Hub. Access all major AI tools in one seamless app, designed to elevate your productivity and streamline your digital experience. Available now on the web at readaloudforme.com and across all your favorite app stores: Apple, Google, and Microsoft. "Read Aloud For Me" brings the future of AI directly to your fingertips, merging convenience with innovation. Whether for work, education, or personal enhancement, our app is your gateway to the most advanced AI technologies. Download today and transform the way you interact with AI tools.
Read Aloud For Me – AI Dashboard: All-in-One AI Tool Hub

A daily chronicle of AI Innovations: March 31st 2024: Generative AI develops potential new drugs for antibiotic-resistant bacteria; South Korean ‘artificial sun’ hits record 100M degrees for 100 seconds; Summary of the key points about OpenAI’s relationship with Dubai and the UAE; Deepmind did not originally see LLMs and the transformer as a path to AGI. Fascinating article.

Generative AI develops potential new drugs for antibiotic-resistant bacteria

Stanford Medicine researchers devise a new artificial intelligence model, SyntheMol, which creates recipes for chemists to synthesize the drugs in the lab.

With nearly 5 million deaths linked to antibiotic resistance globally every year, new ways to combat resistant bacterial strains are urgently needed.

Researchers at Stanford Medicine and McMaster University are tackling this problem with generative artificial intelligence. A new model, dubbed SyntheMol (for synthesizing molecules), created structures and chemical recipes for six novel drugs aimed at killing resistant strains of Acinetobacter baumannii, one of the leading pathogens responsible for antibacterial resistance-related deaths.

The researchers described their model and experimental validation of these new compounds in a study published March 22 in the journal Nature Machine Intelligence.

“There’s a huge public health need to develop new antibiotics quickly,” said James Zou, PhD, an associate professor of biomedical data science and co-senior author on the study. “Our hypothesis was that there are a lot of potential molecules out there that could be effective drugs, but we haven’t made or tested them yet. That’s why we wanted to use AI to design entirely new molecules that have never been seen in nature.”

Source

South Korean ‘artificial sun’ hits record 100M degrees for 100 seconds

For the first time, the Korea Institute of Fusion Energy’s (KFE) Korea Superconducting Tokamak Advanced Research (KSTAR) fusion reactor has reached temperatures seven times that of the Sun’s core.

Achieved during testing between December 2023 and February 2024, this sets a new record for the fusion reactor project.

KSTAR, the researchers behind the reactor report, managed to maintain temperatures of 212 degrees Fahrenheit (100 million degrees Celsius) for 48 seconds. For reference, the temperature of the core of our Sun is 27 million degrees Fahrenheit (15 million degrees Celsius).

Source

Gemini 1.5 Pro on Vertex AI is available for everyone as an experimental release

I think this one has flown under the radar: Gemini 1.5 Pro is available as Experimental on Vertex AI, for everyone, UI only for now (no API yet). In us-central1.

You find it under Vertex AI –> Multimodal. It’s called Gemini Experimental.

API, more features and so on are coming as we approach Google Cloud Next (April 9-11).

OpenAI Relationships

“Summary of the key points about OpenAI’s relationship with Dubai and the UAE”

OpenAI’s Partnership with G42

  • In October 2023, G42, a leading UAE-based technology holding group, announced a partnership with OpenAI to deliver advanced AI solutions to the UAE and regional markets.
  • The partnership will focus on leveraging OpenAI’s generative AI models in domains where G42 has deep expertise, including financial services, energy, healthcare, and public services.
  • G42 will prioritize its substantial AI infrastructure capacity to support OpenAI’s local and regional inferencing on Microsoft Azure data centers.
  • Sam Altman, CEO of OpenAI, stated that the collaboration with G42 aims to empower businesses and communities with effective solutions that resonate with the nuances of the region.

Altman’s Vision for the UAE as an AI Sandbox

  • During a virtual appearance at the World Governments Summit, Altman suggested that the UAE could serve as the world’s “regulatory sandbox” to test AI technologies and later spearhead global rules limiting their use.
  • Altman believes the UAE is well-positioned to be a leader in discussions about unified global policies to rein in future advances in AI.
  • The UAE has invested heavily in AI and made it a key policy consideration.

Altman’s Pursuit of Trillions in Funding for AI Chip Manufacturing

  • Altman is reportedly in talks with investors, including the UAE, to raise $5-7 trillion for AI chip manufacturing to address the scarcity of GPUs crucial for training and running large language models.
  • As part of the talks, Altman is pitching a partnership between OpenAI, various investors, chip makers, and power providers to build chip foundries that would be run by existing chip makers, with OpenAI agreeing to be a significant customer.

In summary, OpenAI’s partnership with G42 aims to expand AI capabilities in the UAE and the Middle East, with Altman envisioning the UAE as a potential global AI sandbox.

Deepmind did not originally see LLMs and the transformer as a path to AGI. Fascinating article.

https://www.bigtechnology.com/p/can-demis-hassabis-save-google

It’s a very long article so I’ll post the relevant snippets. But basically it seems that Google was late to the LLM game because Demis Hassabis was 100% focused on AGI and did not see LLM’s as a path toward AGI. Perhaps now he sees it as a potential path, but it’s probably possible that he is just now focusing on LLM’s so that Google does not get too far behind in the generative AI race. But his ultimate goal and obsession is to create AGI that can solve real problems like diseases.

Within DeepMind, generative models weren’t taken seriously enough, according to those inside, perhaps because they didn’t align with Hassabis’s AGI priority, and weren’t close to reinforcement learning. Whatever the rationale, DeepMind fell behind in a key area.”

“‘We’ve always had amazing frontier work on self-supervised and deep learning,’ Hassabis tells me. ‘But maybe the engineering and scaling component — that we could’ve done harder and earlier. And obviously we’re doing that completely now.'”

“Kulkarni, the ex-DeepMind engineer, believes generative models were not respected at the time across the AI field, and simply hadn’t show enough promise to merit investment. ‘Someone taking the counter-bet had to pursue that path,’ he says. ‘That’s what OpenAI did.'”

“Ironically, a breakthrough within Google — called the transformer model — led to the real leap. OpenAI used transformers to build its GPT models, which eventually powered ChatGPT. Its generative ‘large language’ models employed a form of training called “self-supervised learning,” focused on predicting patterns, and not understanding their environments, as AlphaGo did. OpenAI’s generative models were clueless about the physical world they inhabited, making them a dubious path toward human level intelligence, but would still become extremely powerful.”

As DeepMind rejoiced, a serious challenge brewed beneath its nose. Elon Musk and Sam Altman founded OpenAI in 2015, and despite plenty of internal drama, the organization began working on text generation.”

“As OpenAI worked on the counterbet, DeepMind and its AI research counterpart within Google, Google Brain, struggled to communicate. Multiple ex-DeepMind employees tell me their division had a sense of superiority. And it also worked to wall itself off from the Google mothership, perhaps because Google’s product focus could distract from the broader AGI aims. Or perhaps because of simple tribalism. Either way, after inventing the transformer model, Google’s two AI teams didn’t immediately capitalize on it.”

“‘I got in trouble for collaborating on a paper with a Brain because the thought was like, well, why would you collaborate with Brain?’ says one ex-DeepMind engineer. ‘Why wouldn’t you just work within DeepMind itself?'”

“Then, a few months later, OpenAI released ChatGPT.” “At first, ChatGPT was a curiosity. The OpenAI chatbot showed up on the scene in late 2022 and publications tried to wrap their heads around its significance. […] Within Google, the product felt familiar to LaMDA, a generative AI chatbot the company had run internally — and even convinced one employee it was sentient — but never released. When ChatGPT became the fastest growing consumer product in history, and seemed like it could be useful for search queries, Google realized it had a problem on its hands.”

OpenAI reveals Voice Engine, but won’t yet publicly release the risky AI voice-cloning technology

OpenAI has released VoiceEngine, a voice-cloning tool. The company claims that it can recreate a person’s voice with just 15 seconds of recording of that person talking.

Source

A museum is using AI to let visitors chat with World War II survivors. [Source]

Meta to Add AI to Ray-Ban Smart Glasses. [Source]

Demis Hassabis, CEO and one of three founders of Google’s artificial intelligence (AI) subsidiary DeepMind, has been awarded a knighthood in the U.K. for “services to artificial intelligence.” [Source]

A daily chronicle of AI Innovations: March 30th, 2024: 🤯 Microsoft and OpenAI to build $100 billion AI supercomputer ‘Stargate’; 🗣 OpenAI unveils voice-cloning tool; 📈 Amazon’s AI team faces pressure to outperform Anthropic’s Claude models by mid-year;  🚫 Microsoft Copilot has been blocked on all Congress-owned devices

Microsoft and OpenAI to build $100 billion AI supercomputer ‘Stargate’

  • OpenAI and Microsoft are working on a $100 billion project to build an AI supercomputer named ‘Stargate’ in the U.S.
  • The supercomputer will house millions of GPUs and could cost over $115 billion.
  • Stargate is part of a series of datacenter projects planned by the two companies, with the goal of having it operational by 2028.
  • Microsoft will fund the datacenter, which is expected to be 100 times more costly than current operating centers.
  • The supercomputer is being built in phases, with Stargate being a phase 5 system.
  • Challenges include designing novel cooling systems and considering alternative power sources like nuclear energy.
  • OpenAI aims to move away from Nvidia’s technology and use Ethernet cables instead of InfiniBand cables.
  • Details about the location and structure of the supercomputer are still being finalized.
  • Both companies are investing heavily in AI infrastructure to advance the capabilities of AI technology.
  • Microsoft’s partnership with OpenAI is expected to deepen with the development of projects like Stargate.

Source


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)
  • Microsoft and OpenAI are reportedly collaborating on a significant project to create a U.S.-based datacenter for an AI supercomputer named “Stargate,” estimated to cost over $115 billion and utilize millions of GPUs.
  • The supercomputer aims to be the largest among the datacenters planned by the two companies within the next six years, with Microsoft covering the costs and aiming for a launch by 2028.
  • The project, considered to be in phase 5 of development, requires innovative solutions for power, cooling, and hardware efficiency, including a possible shift away from relying on Nvidia’s InfiniBand in favor of Ethernet cables.
  • Source

🗣 OpenAI unveils voice-cloning tool

  • OpenAI has developed a text-to-voice generation platform named Voice Engine, capable of creating a synthetic voice from just a 15-second voice clip.
  • The platform is in limited access, serving entities like the Age of Learning and Livox, and is being used for applications from education to healthcare.
  • With concerns around ethical use, OpenAI has implemented usage policies, requiring informed consent and watermarking audio to ensure transparency and traceability.
  • Source

📈 Amazon’s AI team faces pressure to outperform Anthropic’s Claude models by mid-year

  • Amazon has invested $4 billion in AI startup Anthropic, but is also developing a competing large-scale language model called Olympus.
  • Olympus is supposed to surpass Anthropic’s latest Claude model by the middle of the year and has “hundreds of billions of parameters.”
  • So far, Amazon has had no success with its own language models. Employees are unhappy with Olympus’ development time and are considering switching to Anthropic’s models.
  • Source

🚫 Microsoft Copilot has been blocked on all Congress-owned devices

  • The US House of Representatives has banned its staff from using Microsoft’s AI chatbot Copilot due to cybersecurity concerns over potential data leaks.
  • Microsoft plans to remove Copilot from all House devices and is developing a government-specific version aimed at meeting federal security standards.
  • The ban specifically targets the commercial version of Copilot, with the House open to reassessing a government-approved version upon its release.
  • Source

Official NYC chatbot is encouraging small businesses to break the law.LINK

ChatGPT’s responses now include source references but for paid users.LINK

Next-generation AI semiconductor devices mimic the human brain.LINK

Voicecraft: I’ve never been more impressed in my entire life !

Voicecraft
Voicecraft

The maintainers of Voicecraft published the weights of the model earlier today, and the first results I get are incredible.

Here’s only one example, it’s not the best, but it’s not cherry-picked, and it’s still better than anything I’ve ever gotten my hands on !

Here’s the Github repository for those interested: https://github.com/jasonppy/VoiceCraft

A daily chronicle of AI Innovations: March 29th, 2024: 💥 Apple files lawsuit against former engineer for leaking details of projects he wanted to kill; Apple files lawsuit against former engineer for leaking details of projects he wanted to kill; Microsoft tackles Gen AI risks with new Azure AI tools; AI21 Labs’ Jamba triples AI throughput ; Google DeepMind’s AI fact-checker outperforms humans ; X’s Grok gets a major upgrade; Lightning AI partners with Nvidia to launch Thunder AI compiler

💥 Apple files lawsuit against former engineer for leaking details of projects he wanted to kill

  • Apple has filed a lawsuit against former employee Andrew Aude for leaking confidential information about products like the Vision Pro and Journal app to journalists and competitors, motivated by his desire to “kill” products and features he disagreed with.
  • Aude, who joined Apple in 2016, is accused of sharing sensitive details via encrypted messages and meetings, including over 10,000 text messages to a journalist from The Information.
  • The lawsuit seeks damages, the return of bonuses and stock options, and a restraining order against Aude for disclosing any more of Apple’s confidential information.
  • Source

👮‍♂️ Microsoft launches tools to try and stop people messing with chatbots

  • Microsoft has introduced a new set of tools in Azure to enhance the safety and security of generative AI applications, especially chatbots, aiming to counter risks like abusive content and prompt injections.
  • The suite includes features for real-time monitoring and protection against sophisticated threats, leveraging advanced machine learning to prevent direct and indirect prompt attacks.
  • These developments reflect Microsoft’s ongoing commitment to responsible AI usage, fueled by its significant investment in OpenAI and intended to address the security and reliability concerns of corporate leaders.
  • Source

AI21 Labs’ Jamba triples AI throughput

AI21 Labs has released Jamba, the first-ever production-grade AI model based on the Mamba architecture. This new architecture combines the strengths of both traditional Transformer models and the Mamba SSM, resulting in a model that is both powerful and efficient. Jamba boasts a large context window of 256K tokens, while still fitting on a single GPU.

AI21 Labs’ Jamba triples AI throughput
AI21 Labs’ Jamba triples AI throughput

Jamba’s hybrid architecture, composed of Transformer, Mamba, and mixture-of-experts (MoE) layers, optimizes for memory, throughput, and performance simultaneously.

The model has demonstrated remarkable results on various benchmarks, matching or outperforming state-of-the-art models in its size class. Jamba is being released with open weights under Apache 2.0 license and will be accessible from the NVIDIA API catalog.

Why does this matter?

Jamba’s hybrid architecture makes it the only model capable of processing 240k tokens on a single GPU. This could make AI tasks like machine translation and document analysis much faster and cheaper, without requiring extensive computing resources.

Source

Google DeepMind’s AI fact-checker outperforms humans

Google DeepMind has developed an AI system called Search-Augmented Factuality Evaluator (SAFE) that can evaluate the accuracy of information generated by large language models more effectively than human fact-checkers. In a study, SAFE matched human ratings 72% of the time and was correct in 76% of disagreements with humans.

Google DeepMind's AI fact-checker outperforms humans
Google DeepMind’s AI fact-checker outperforms humans

While some experts question the use of “superhuman” to describe SAFE’s performance, arguing for benchmarking against expert fact-checkers, the system’s cost-effectiveness is undeniable, being 20 times cheaper than human fact-checkers.

Why does this matter?

As language models become more powerful and widely used, SAFE could combat misinformation and ensure the accuracy of AI-generated content. SAFE’s efficiency could be a game-changer for consumers relying on AI for tasks like research and content creation.

Source

X’s Grok gets a major upgrade

X.ai, Elon Musk’s AI startup, has introduced Grok-1.5, an upgraded AI model for their Grok chatbot. This new version enhances reasoning skills, especially in coding and math tasks, and expands its capacity to handle longer and more complex inputs with a 128,000-token context window.

X’s Grok gets a major upgrade
X’s Grok gets a major upgrade

Grok chatbots are known for their ability to discuss controversial topics with a rebellious touch. The improved model will first be tested by early users on X, with plans for wider availability later. This release follows the open-sourcing of Grok-1 and the inclusion of the chatbot in X’s $8-per-month Premium plan.

Why does this matter?

This is significant because Grok-1.5 represents an advancement in AI assistants, potentially offering improved help with complex tasks and better understanding of user intent through its larger context window and real-time data ability. This could impact how people interact with chatbots in the future, making them more helpful and reliable.

Source

What Else Is Happening in AI on March 29th, 2024❗

🛡️Microsoft tackles Gen AI risks with new Azure AI tools

Microsoft has launched new Azure AI tools to address the safety and reliability risks associated with generative AI. The tools, currently in preview, aim to prevent prompt injection attacks, hallucinations, and the generation of personal or harmful content. The offerings include Prompt Shields, prebuilt templates for safety-centric system messages, and Groundedness Detection.  (Link)

🤝Lightning AI partners with Nvidia to launch Thunder AI compiler

Lightning AI, in collaboration with Nvidia, has launched Thunder, an open-source compiler for PyTorch, to speed up AI model training by optimizing GPU usage. The company claims that Thunder can achieve up to a 40% speed-up for training large language models compared to unoptimized code. (Link)

🥊SambaNova’s new AI model beats Databricks’ DBRX

SambaNova Systems’ Samba-CoE v0.2 Large Language Model outperforms competitors like Databricks’ DBRX, MistralAI’s Mixtral-8x7B, and xAI’s Grok-1. With 330 tokens per second using only 8 sockets, Samba-CoE v0.2 demonstrates remarkable speed and efficiency without sacrificing precision. (Link)

🌍Google.org launches Accelerator to empower nonprofits with Gen AI

Google.org has announced a six-month accelerator program to support 21 nonprofits in leveraging generative AI for social impact. The program provides funding, mentorship, and technical training to help organizations develop AI-powered tools in areas such as climate, health, education, and economic opportunity, aiming to make AI more accessible and impactful. (Link)

📱Pixel 8 to get on-device AI features powered by Gemini Nano

Ace the Microsoft Azure Fundamentals AZ-900 Certification Exam: Pass the Azure Fundamentals Exam with Ease

Google is set to introduce on-device AI features like recording summaries and smart replies on the Pixel 8, powered by its small-sized Gemini Nano model. The features will be available as a developer preview in the next Pixel feature drop, marking a shift from Google’s primarily cloud-based AI approach. (Link)

A daily chronicle of AI Innovations: March 28th, 2024: ⚡ DBRX becomes world’s most powerful open-source LLM 🏆 Claude 3 Opus crowned the top user-rated chatbot, beating OpenAI’s GPT-4 💙 Empathy meets AI: Hume AI’s EVI redefines voice interaction

DBRX becomes world’s most powerful open source LLM

Databricks has released DBRX, a family of open-source large language models setting a new standard for performance and efficiency.  The series includes DBRX Base and DBRX Instruct, a fine-tuned version designed for few-turn interactions. Developed by Databricks’ Mosaic AI team and trained using NVIDIA DGX Cloud, these models leverage an optimized mixture-of-experts (MoE) architecture based on the MegaBlocks open-source project. This architecture allows DBRX to achieve up to twice the compute efficiency of other leading LLMs.

DBRX becomes world’s most powerful open source LLM
DBRX becomes world’s most powerful open source LLM

In terms of performance, DBRX outperforms open-source models like Llama 2 70B, Mixtral-8x7B, and Grok-1 on industry benchmarks for language understanding, programming, and math. It also surpasses GPT-3.5 on most of these benchmarks, although it still lags behind GPT-4. DBRX is available under an open license with some restrictions and can be accessed through GitHub, Hugging Face, and major cloud platforms. Organizations can also leverage DBRX within Databricks’ Data Intelligence Platform.

Why does this matter?

With DBRX, organizations can build and fine-tune powerful proprietary models using their own internal datasets, ensuring full control over their data rights. As a result, DBRX is likely to accelerate the trend of organizations moving away from closed models and embracing open alternatives that offer greater control and customization possibilities.

Source

Claude 3 Opus crowned the top user-rated chatbot, beating OpenAI’s GPT-4

Anthropic’s Claude 3 Opus has overtaken OpenAI’s GPT-4 to become the top-rated chatbot on the Chatbot Arena leaderboard. This marks the first time in approximately a year since GPT-4’s release that another language model has surpassed it in this benchmark, which ranks models based on user preferences in randomized head-to-head comparisons. Anthropic’s cheaper Haiku and mid-range Sonnet models also perform impressively, coming close to the original GPT-4’s capabilities at a significantly lower cost.

Claude 3 Opus crowned the top user-rated chatbot, beating OpenAI’s GPT-4
Claude 3 Opus crowned the top user-rated chatbot, beating OpenAI’s GPT-4

While OpenAI still dominates the market, especially among regular users with ChatGPT, this development and recent leadership changes at OpenAI have helped Anthropic gain ground. However, OpenAI is rumored to be preparing to launch an even more advanced “GPT-4.5” or “GPT-5” model as soon as this summer, which CEO Sam Altman has teased will be “amazing,” potentially allowing them to retake the lead from Anthropic’s Claude 3 Opus.

Why does this matter?

Claude’s rise to the top of the Chatbot Arena leaderboard shows that OpenAI is not invincible and will face stiff competition in the battle for AI supremacy. With well-resourced challengers like Anthropic and Google, OpenAI will need to move fast and innovate boldly to maintain its top position. Ultimately, this rivalry will benefit everyone as it catalyzes the development of more powerful, capable, and hopefully beneficial AI systems that can help solve humanity’s major challenges.

Source

If you are looking for an all-in-one solution to help you prepare for the AWS Cloud Practitioner Certification Exam, look no further than this AWS Cloud Practitioner CCP CLF-C02 book

Empathy meets AI: Hume AI’s EVI redefines voice interaction

In a significant development for the AI community, Hume AI has introduced a new conversational AI called Empathic Voice Interface (EVI). What sets EVI apart from other voice interfaces is its ability to understand and respond to the user’s tone of voice, adding unprecedented emotional intelligence to the interaction. By adapting its language and responses based on the user’s expressions, EVI creates a more human-like experience, blurring the lines between artificial and emotional intelligence.

EVI’s empathic capabilities extend beyond just understanding tone. It can accurately detect the end of a conversation turn, handle interruptions seamlessly, and even learn from user reactions to improve over time. These features, along with its fast and reliable transcription and text-to-speech capabilities, make EVI a highly adaptable tool for various applications. Developers can easily integrate EVI into their projects using Hume’s API, which will be publicly available in April.

Why does this matter?

Emotionally intelligent AI can be revolutionary for industries like healthcare and use cases like customer support, where empathy and emotional understanding are crucial. But we must also consider potential risks, such as overreliance on AI for emotional support or the possibility of AI systems influencing users’ emotions in unintended ways. If developed and implemented ethically, emotionally intelligent AI can greatly enhance how we interact with and benefit from AI technologies in our daily lives.

Source

What Else Is Happening in AI on March 28th, 2024❗

💰 OpenAI launches revenue sharing program for GPT Store builders

OpenAI is experimenting with sharing revenue with builders who create successful apps using GPT in OpenAI’s GPT Store. The goal is to incentivize creativity and collaboration by rewarding builders for their impact on an ecosystem OpenAI is testing so they can make it easy for anyone to build and monetize AI-powered apps. (Link)

🛍️ Google introduces new shopping features to refine searches

Google is rolling out new shopping features that allow users to refine their searches and find items they like more easily. The Style Recommendations feature lets shoppers rate items in their searches, helping Google pick up on their preferences. Users can also specify their favorite brands to instantly bring up more apparel from those selections.  (Link)

🗣️ rabbit’s r1 device gets ultra-realistic voice powered by ElevenLabs

ElevenLabs has partnered with rabbit to integrate its high-quality, low-latency voice AI into rabbit’s r1 AI companion device. The collaboration aims to make the user experience with r1 more natural and intuitive by allowing users to interact with the device using voice commands. (Link)

💸 AI startup Hume raises $50M to build emotionally intelligent conversational AI

AI startup Hume has raised $50 million in a Series B funding round, valuing the company at $219 million. Hume’s AI technology can detect over 24 distinct emotional expressions in human speech and generate appropriate responses. The startup’s AI has been integrated into applications across healthcare, customer service, and productivity, with the goal of providing more context and empathy in AI interactions. (Link)

💻 Lenovo launches AI-enhanced PCs in a push for innovation and differentiation

Lenovo revealed a new lineup of AI-powered PCs and laptops at its Innovate event in Bangkok, Thailand. The company showcased the dual-screen Yoga Book 9i, Yoga Pro 9i with an AI chip for performance optimization and AI-enhanced Legion gaming laptops. Lenovo hopes to differentiate itself in the crowded PC market and revive excitement with these AI-driven innovations. (Link)

Study shows ChatGPT can produce medical record notes 10 times faster than doctors without compromising quality

The AI model ChatGPT can write administrative medical notes up to 10 times faster than doctors without compromising quality. This is according to a study conducted by researchers at Uppsala University Hospital and Uppsala University in collaboration with Danderyd Hospital and the University Hospital of Basel, Switzerland. The research is published in the journal Acta Orthopaedica.

Source

Microsoft Copilot AI will soon run locally on PCs

Microsoft’s Copilot AI service is set to run locally on PCs, Intel told Tom’s Hardware. The company also said that next-gen AI PCs would require built-in neural processing units (NPUs) with over 40 TOPS (trillion operations per second) of power — beyond the capabilities of any consumer processor on the market.

Intel said that the AI PCs would be able to run “more elements of Copilot” locally. Currently, Copilot runs nearly everything in the cloud, even small requests. That creates a fair amount of lag that’s fine for larger jobs, but not ideal for smaller jobs. Adding local compute capability would decrease that lag, while potentially improving performance and privacy as well.

Microsoft was previously rumored to require 40 TOPS on next-gen AI PCs (along with a modest 16GB of RAM). Right now, Windows doesn’t make much use of NPUs, apart from running video effects like background blurring for Surface Studio webcams. ChromeOS and macOS both use NPU power for more video and audio processing features, though, along with OCR, translation, live transcription and more, Ars Technica noted.

Source

A daily chronicle of AI Innovations: March 27th, 2024: 🔥 Microsoft study reveals the 11 by 11 tipping point for AI adoption 🤖 A16z spotlights the rise of generative AI in enterprises 🚨 Gaussian Frosting revolutionizes surface reconstruction in 3D modeling 🤖OpenAI unveils exciting upcoming features for GPT-4 and DALL-E 3 🤖 Adobe unveils GenStudio: AI-powered ad creation platform

Microsoft study reveals the 11 by 11 tipping point for AI adoption

Microsoft’s study on AI adoption in the workplace revealed the “11-by-11 tipping point,” where users start seeing AI’s value by saving 11 minutes daily. The study involved 1,300 Copilot for Microsoft 365 users and showed that 11 minutes of time savings is enough for most people to find AI useful.

Microsoft study reveals the 11 by 11 tipping point for AI adoption
Microsoft study reveals the 11 by 11 tipping point for AI adoption

Over 11 weeks, users reported improved productivity, work enjoyment, work-life balance, and fewer meetings. This “11-by-11 tipping point” signifies the time it takes for individuals to experience AI’s benefits in their work fully.

Why does it matter?

The study offers insights for organizations aiming to drive AI adoption among their employees. Businesses can focus on identifying specific use cases that deliver immediate benefits like time and cost savings. It will help organizations encourage employees to embrace AI, increasing productivity and improving work experiences.

Source

A16z spotlights the rise of generative AI in enterprises

A groundbreaking report by the influential tech firm a16z  unveils the rapid integration of generative AI technologies within the corporate sphere. The report highlights essential considerations for business leaders to harness generative AI effectively. It covers resource allocation, model selection, and innovative use cases, providing a strategic roadmap for enterprises.

A16z spotlights the rise of generative AI in enterprises
A16z spotlights the rise of generative AI in enterprises

An increased financial commitment from businesses marks the adoption of generative AI. Industry leaders are tripling their investments in AI technologies, emphasizing the pivotal role of generative AI in driving innovation and efficiency.

A16z spotlights the rise of generative AI in enterprises
A16z spotlights the rise of generative AI in enterprises

The shift towards integrating AI into core operations is evident. There is a focus on measuring productivity gains and cost savings and quantifying impact on key business metrics.

Why does it matter?

The increasing budgets allocated to generative AI signal its strategic importance in driving innovation and productivity in enterprises. This highlights AI’s transformative potential to provide a competitive edge and unlock new opportunities. Generative AI can revolutionize various business operations and help gain valuable insights by leveraging diverse data types.

Source

Gaussian Frosting revolutionizes surface reconstruction in 3D modeling

At the international conference on computer vision, researchers presented a new method to improve surface reconstruction using Gaussian Frosting. This technique automates the adjustment of Poisson surface reconstruction hyperparameters, resulting in significantly improved mesh reconstruction.

Gaussian Frosting revolutionalizes surface reconstruction in 3D modeling
Gaussian Frosting revolutionalizes surface reconstruction in 3D modeling

The method showcases the potential for scaling up mesh reconstruction while preserving intricate details and opens up possibilities for advanced geometry and texture editing. This work marks a significant step forward in surface reconstruction methods, promising advancements in 3D modeling and visualization techniques.

Why does it matter?

The new method demonstrates how AI enhances surface reconstruction techniques, improving mesh quality and enabling advanced editing in 3D modeling. This has significant implications for revolutionizing how 3D models are created, edited, and visualized across various industries.

Source

AIs can now learn and talk with each other like humans do.

This seems an important step toward AGI and vastly improved productivity.

“Once these tasks had been learned, the network was able to describe them to a second network — a copy of the first — so that it could reproduce them. To our knowledge, this is the first time that two AIs have been able to talk to each other in a purely linguistic way,’’ said lead author of the paper Alexandre Pouget, leader of the Geneva University Neurocenter, in a statement.”

“While AI-powered chatbots can interpret linguistic instructions to generate an image or text, they can’t translate written or verbal instructions into physical actions, let alone explain the instructions to another AI.

However, by simulating the areas of the human brain responsible for language perception, interpretation and instructions-based actions, the researchers created an AI with human-like learning and communication skills.”

Source

What Else Is Happening in AI on March 27th, 2024❗

🤖 Adobe unveils GenStudio: AI-powered ad creation platform

Adobe introduced GenStudio, an AI-powered ad creation platform, during its Summit event. GenStudio is a centralized hub for promotional campaigns, offering brand kits, copy guidance, and preapproved assets. It also provides generative AI-powered tools for generating backgrounds and ensuring brand consistency. Users can quickly create ads for email and social media platforms like Facebook, Instagram, and LinkedIn. (Link)

🧑‍💼Airtable introduces AI summarization for enhanced productivity

Airtable has introduced Airtable AI, which provides generative AI summarization, categorization, and translation to users. This feature allows quick insights and understanding of information within workspaces, enabling easy sharing of valuable insights with teams. Airtable AI automatically applies categories and tags to information, routes action items to the relevant team, and generates emails or social posts with a single button tap. (Link)

🤝Microsoft Teams enhances Copilot AI features for improved collaboration

Microsoft is introducing smarter Copilot AI features in Microsoft Teams to enhance collaboration and productivity. The updates include new ways to invoke the assistant during meeting chats and summaries, making it easier to catch up on missed meetings by combining spoken transcripts and written chats into a single view. Microsoft is launching new hybrid meeting features, such as automatic camera switching for remote participants and speaker recognition for accurate transcripts. (Link)

🤖OpenAI unveils exciting upcoming features for GPT-4 and DALL-E 3

OpenAI is preparing to introduce new features for its GPT-4 and DALL-E 3 models. For GPT-4, OpenAI plans to remove the message limit, implement a Model Tuner Selector, and allow users to upgrade responses from GPT-3.5 to GPT-4 with a simple button push. On the DALL-E 3 front, OpenAI is working on an image editor with inpainting functionality. These upcoming features demonstrate OpenAI’s commitment to advancing AI capabilities. (Link)

🔍Apple Chooses Baidu’s AI for iPhone 16 in China

Apple has reportedly chosen Baidu to provide AI technology for its upcoming iPhone 16 and other devices in China. This decision comes as Apple faces challenges due to stagnation in iPhone innovation and competition from Huawei. Baidu’s Ernie Bot will be included in the Chinese version of the iPhone 16, Mac OS, and iOS 18. Despite discussions with Alibaba Group Holding and a Tsinghua University AI startup, Apple selected Baidu’s AI technology for compliance. (Link)

Meta CEO, Mark Zuckerberg, is directly recruiting AI talent from Google’s DeepMind with personalized emails.

Meta CEO, Mark Zuckerberg, is attempting to recruit top AI talent from Google’s DeepMind (their AI research unit). Personalised emails, from Zuckerberg himself, have been sent to a few of their top researchers, according to a report from The Information, which cited individuals that had seen the messages. In addition to this, the researchers are being hired without having to do any interviews, and, a previous policy which Meta had in place – to not offer higher offers to candidates with competing job offers – has been relaxed.

Zuckerberg appears to be on a hiring spree to build Meta into a position of being a dominant player in the AI space.

OpenAI’s Sora Takes About 12 Minutes to Generate 1 Minute Video on NVIDIA H100. Source.

Apple on Tuesday announced that its annual developers conference, WWDC, will take place June 10 through June 14. Source.

Elon Musk says all Premium subscribers on X will gain access to AI chatbot Grok this week. Source.

Intel unveils AI PC program for software developers and hardware vendors. Source.

London-made HIV injection has potential to cure millions worldwide

Source

A daily chronicle of AI Innovations: March 26th, 2024 : 🔥 Zoom launches all-in-one modern AI collab platform; 🤖 Stability AI launches instruction-tuned LLM; 🚨 Stability AI CEO resigns to focus on decentralized AI; 🔍 WhatsApp to integrate Meta AI directly into its search bar; 🥊 Google, Intel, and Qualcomm challenge Nvidia’s dominance in AI; 🎬 OpenAI pitches Sora to Hollywood studios

Zoom launches all-in-one modern AI collab platform

Zoom launched Zoom Workplace, an AI collaboration platform that integrates many tools to improve teamwork and productivity. With over 40 new features, including AI Companion updates for Zoom Phone, Team Chat, Events, and Contact Center, as well as the introduction of Ask AI Companion, Zoom Workplace simplifies workflows within a familiar interface.

The platform offers customization options, meeting features, and improved collaboration tools across Zoom’s ecosystem. Zoom Business Services, integrated with Zoom Workplace, offers AI-driven marketing, customer service, and sales solutions. It expands digital communication channels and provides real-time insights for better agent management.

Why does this matter?

This intelligent platform will increase productivity by automating tasks, summarizing interactions, and personalizing user experiences. This move positions Zoom as a frontrunner in the race to integrate AI into everyday work tools, which will reshape how teams communicate and collaborate.

Source

Stability AI launches instruction-tuned LLM

Stability AI has introduced Stable Code Instruct 3B, a new instruction-tuned large language model. It can handle various software development tasks, such as code completion, generation, translation, and explanation, as well as creating database queries with simple instructions.

Stable Code Instruct 3B claims to outperform rival models like CodeLlama 7B Instruct and DeepSeek-Coder Instruct 1.3B in terms of accuracy, understanding natural language instructions, and handling diverse programming languages. The model is accessible for commercial use with a Stability AI Membership, while its weights are freely available on Hugging Face for non-commercial projects.

Why does this matter?

This model simplifies development workflows and complex tasks by providing contextual code completion, translation, and explanations. Businesses can prototype, iterate and ship software products faster thanks to its high performance and low hardware requirements.

Source

Stability AI CEO resigns because of centralized AI

  • Stability AI CEO Emad Mostaque steps down to focus on decentralized AI, advocating for transparent governance in the industry.

  • Mostaque’s departure follows the appointment of interim co-CEOs Shan Shan Wong and Christian Laforte.

  • The startup, known for its image generation tool, faced challenges including talent loss and financial struggles.

  • Mostaque emphasized the importance of generative AI R&D over revenue growth and highlighted the potential economic value of open models in regulated industries.

  • The AI industry witnessed significant changes with Inflection AI co-founders joining Microsoft after raising $1.5 billion.

Source

Estimating Sora’s power requirements

Quoting the compute estimates of Sora from the factorial funds blog

Estimating Sora's power requirements
Estimating Sora’s power requirements

A 15% penetration of Sora for videos with realistic video generation demand and utilization will require about 720k Nvidia H100 GPUs. Each H100 requires about 700 Watts of power supply.

720,000 x 700 = 504 Megawatts.

By comparison, even the largest ever fully solar powered plan in America (Ivanpah Solar Power Facility) produces about 377 Megawats.

While these power requirements can be met with other options like nuclear plants and even coal/hydro plants of big sizes … are we really entering the power game for electricity ?

( it is currently a power game on compute)

What Else Is Happening in AI on March 26th, 2024❗

💬 The Financial Times has introduced Ask FT, a new GenAI chatbot

It provides curated, natural-language responses to queries about recent events and broader topics covered by the FT. Ask FT is powered by Anthropic’s Claude and is available to a selected group of subscribers as it is under testing. (Link)

🔍 WhatsApp to integrate Meta AI directly into its search bar

The latest Android WhatsApp beta update will embed Meta AI directly into the search bar. This feature will allow users to type queries into the search bar and receive instant AI-powered responses without creating a separate Meta AI chat. The update will also allow users to interact with Meta AI even if they choose to hide the shortcut. (Link)

🥊 Google, Intel, and Qualcomm challenge Nvidia’s dominance in AI 

Qualcomm, Google, and Intel are targeting NVIDIA’s software platforms like CUDA. They plan to create open-source tools compatible with multiple AI accelerator chips through the UXL Foundation. Companies are investing over $4 billion in startups developing AI software to loosen NVIDIA’s grip on the field. (Link)

🤖 Apple takes a multi-vendor approach for generative AI in iOS 18

Apple is reportedly in talks with Alphabet, OpenAI, and Anthropic to integrate generative AI capabilities from multiple vendors into iOS 18. This multi-vendor approach aligns with Apple’s efforts to balance advanced AI features with privacy considerations, which are expected to be detailed at WWDC 2024 during the iOS 18 launch. (Link)

🎬 OpenAI pitches Sora to Hollywood studios

OpenAI is actively engaging with Hollywood studios, directors, and talent agencies to integrate Sora into the entertainment industry. The startup has scheduled meetings in Los Angeles to showcase Sora’s capabilities and encourage partnerships, with CEO Sam Altman attending events during the Oscars weekend. (Link)

LLM providers charge you per token, but their tokens are not always comparable. So if you are putting Python code through GPT-4 and Claude 3, it would cost you 25% more tokens to do so with Claude, due to difference in their tokenisers (note: this is different to cost per token, it just means you will have more tokens to pay for)

Some observations:
– OpenAI’s GPT-4 & 3.5 tokeniser is the most efficient for English and Python
– Gemini absolutely demolishes the competition in the three languages I tested: French (-11%), Chinese (-43%) and Hebrew (-54%)
– If your use cases is non-English, really worth looking at Gemini models – the difference in cost will likely be very noticeable
– Llama 2 ranked at the bottom of all of my tests
– Mistral was kind of disappointing on French (+16% worse than GPT), the reason why I picked French was that I assumed they’d do better

Methodology notes:
– The study will be limited, I only compared 7 individual bits of text/code – so results in practice would vary
– I have used this tokeniser playground (https://huggingface.co/spaces/Xenova/the-tokenizer-playground) for GPT, Mistral and Llama. I found it to be inaccurate (or old?) for Claude 3 and they didn’t have Gemini, so I did these separately
– Tokens are only part of the puzzle, more efficient tokenisation won’t necessarily mean better performance or overall lower cost
– If you want to learn about tokenisers, I recommend watching this video from Andrej Karpathy, even the first 10-20 minutes will be really worth your time https://www.youtube.com/watch?v=zduSFxRajkE

No alt text provided for this image

Source: Peter Gostev

A daily chronicle of AI Innovations: March 25th, 2024 : 🤝 Apple could partner with OpenAI, Gemini, Anthropic; 🤖 Chatbots more likely to change your mind than another human, study says; 🤖 Chatbots more likely to change your mind than another human, study says; Verbal Reasoning Test – Opus is better than 93% of people, Gemini 1.5 Pro 59%, GPT-4 Turbo only 36%; Apple’s Tim Cook says AI essential tool for businesses to reduce carbon footprint; Suno V3: Song-on-demand AI is getting insanely good; The first patient with a Neuralink brain-computer implant played Nintendo’s Mario Kart video game with his mind in an impressive new demo video

🤝 Apple could partner with OpenAI, Gemini, Anthropic

  • Apple is discussing with Alphabet, OpenAI, Anthropic, and potentially Baidu to integrate generative AI into iOS 18, considering multiple partners rather than a single one.
  • The collaboration could lead to a model where iPhone users might choose their preferred AI provider, akin to selecting a default search engine in a web browser.
  • Reasons for partnering with external AI providers include financial benefits, the possibility to quickly adapt through partnership changes or user preferences, and avoiding the complexities of developing and maintaining cloud-based generative AI in-house.
  • Source

🌐 EU probes Apple, Google, Meta under new digital law 

  • The European Commission has initiated five investigations into Apple, Google, and Meta for potential non-compliance with the Digital Markets Act (DMA), focusing on app store rules, search engine preferencing, and advertisement targeting models.
  • Investigations will also examine Apple’s app distribution fee structure and Amazon’s product preferencing, while Meta is given six months to make Messenger interoperable with other messaging services.
  • Companies may face fines up to 10% of their annual global revenue for DMA non-compliance, with the possibility of increased penalties for repeated infringements.
  • Source

🤖 Chatbots more likely to change your mind than another human, study says

  • A study found that personalized chatbots, such as GPT-4, are more likely to change people’s minds compared to human debaters by using tailored arguments based on personal information.
  • The research conducted by the École Polytechnique Fédérale de Lausanne and the Italian Fondazione Bruno Kessler showed an 81.7 percent increase in agreement when GPT-4 had access to participants’ personal data like age, gender, and race.
  • Concerns were raised about the potential misuse of AI in persuasive technologies, especially with the ability to generate detailed user profiles from online activities, urging online platform operators to counter such strategies.
  • Source

OpenAI CEO’s £142 Million Gamble On Unlocking the Secrets to Longer Life, Altman’s vision of extended lifespans may be achievable

Biotech startup Retro Biosciences is undertaking a one-of-a-kind experiment housed in shipping containers, funded by a $180 (£142.78) million investment by tech leader Sam Altman to increase lifespan.

Altman, the 38-year-old tech heavyweight, has been a significant player in the industry. Despite his young age, Altman took the tech realm by storm with offerings like ChatGPT and Sora. Unsurprisingly, his involvement in these groundbreaking projects has propelled him to a level of influence rivaling Mark Zuckerberg and Elon Musk, who is currently embroiled in a lawsuit with OpenAI.

It is also worth noting that the Altman-led AI startup is reportedly planning to launch its own AI-powered search engine to challenge Google’s search dominance. Altman’s visionary investments in tech giants like Reddit, Stripe, Airbnb, and Instacart propelled him to billionaire status. They cemented his influence as a tech giant who relentlessly pushed the boundaries of the industry’s future.

Source

Nvidia announces AI-powered health care 'agents' that outperform nurses — and cost $9 an hour
Nvidia announces AI-powered health care ‘agents’ that outperform nurses — and cost $9 an hour

Apple researchers explore dropping “Siri” phrase and listening with AI instead

  • Apple researchers are investigating the use of AI to identify when a user is speaking to a device without requiring a trigger phrase like ‘Siri’.

  • A study involved training a large language model using speech and acoustic data to detect patterns indicating the need for assistance from the device.

  • The model showed promising results, outperforming audio-only or text-only models as its size increased.

  • Eliminating the ‘Hey Siri’ prompt could raise concerns about privacy and constant listening by devices.

  • Apple’s handling of audio data has faced scrutiny in the past, leading to policy changes regarding user data and Siri recordings.

Source

Suno V3 can do multiple languages in one song. This one is English, Portuguese, Japanese, and Italian. Incredible.

Beneath the vast sky, where dreams lay rooted deep, Mountains high and valleys wide, secrets they keep. Ground beneath my feet, firm and ever true, Earth, you give us life, in shades of brown and green hue.

Sopra o vento, mensageiro entre o céu e o mar, Carregando sussurros, histórias a contar. Dançam as folhas, em um balé sem fim, Vento, o alento invisível, guiando o destino assim.

火のように、情熱が燃えて、 光と暖かさを私たちに与えてくれる。 夜の暗闇を照らす、勇敢な炎、 生命の力、絶えず変わるゲーム。

Acqua, misteriosa forza che tutto scorre, Nei fiumi, nei mari, la vita che ci offre. Specchio del cielo, in te ci riflettiamo, Acqua, fonte di vita, a te ci affidiamo.

Listen here

OpenAI Heading To Hollywood To Pitch Revolutionary “Sora”

Some of the most important meetings in Hollywood history will take place in the coming week, as OpenAI hits Hollywood to show the potential of its “Sora” software to studios, talent agencies, and media executives.

Bloomberg is reporting that OpenAI wants more filmmakers to become familiar with Sora, the text-to-video generator that potentially could upend the way movies are made.

Soon, Everyone Will Own a Robot, Like a Car or Phone Today. Says Figure AI founder

Brett Adcock, the founder of FigureAI robots, the company that recently released a demo video of its humanoid robot conversing with a human while performing tasks, predicts that everyone will own a robot in the future. “Similar to owning a car or phone today,” he said – hinting at the universal adoption of robots as an essential commodity in the future.

“Every human will own a robot in the future, similar to owning a car/phone today,” said Adcock.

A few months ago, Adcock called 2024 the year of Embodied AI, indicating how the future comprises AI in a body form. With robots learning to perform low-complexity tasks, such as picking trash, placing dishes, and even using the coffee machine, Figure robots are being trained to assist a person with house chores.

Source

WhatsApp to embed Meta AI directly into search bar for instant assistance: Report. 

WhatsApp is on the brink of a transformation in user interaction as it reportedly plans to integrate Meta AI directly into its search bar. This move promises to simplify access to AI assistance within the app, eliminating the need for users to navigate to a separate Meta AI conversation.

WhatsApp to embed Meta AI directly into search bar for instant assistance
WhatsApp to embed Meta AI directly into search bar for instant assistance

Source

How People are really using Gen AI

Top-level themes:

1️⃣ Technical Assistance & Troubleshooting (23%)
2️⃣ Content Creation & Editing (22%)
3️⃣ Personal & Professional Support (17%)
4️⃣ Learning & Education (15%)
5️⃣ Creativity & Recreation (13%)
6️⃣ Research, Analysis & Decision Making (10%)

What users are doing:

✔Generating ideas
✔Specific search
✔Editing text
✔Drafting emails
✔Simple explainers
✔Excel formulas
✔Sampling data

🤔 Do you see AI as a tool to enhance your work, or as a threat that could take over your job?

Source: HBR
Image credit: Filtered

How People are really using Gen AI
How People are really using Gen AI

A daily chronicle of AI Innovations: March 22nd, 2024 : 🤖 Nvidia’s Latte 3D generates text-to-3D in seconds! 💰 Saudi Arabia to invest $40 billion in AI 🚀 Open Interpreter’s 01 Light personal pocket AI agent. 🤖 Microsoft introduces a new Copilot for better productivity.
💡Quiet-STaR: LMs can self-train to think before responding
🤯Neuralink’s first brain chip patient plays chess with his mind

Experience the transformative capabilities of AI with “Read Aloud For Me – AI Dashboard” – your ultimate AI Dashboard and Hub.

Nvidia’s Latte 3D generates text-to-3D in seconds!

NVIDIA introduces Latte3D, facilitating the conversion of text prompts into detailed 3D models in less than a second. Developed by NVIDIA’s Toronto lab, Latte3D sets a new standard in generative AI models with its remarkable blend of speed and precision.

Nvidia’s Latte 3D generates text-to-3D in seconds!
Nvidia’s Latte 3D generates text-to-3D in seconds!

LATTE3D has two stages: first, NVIDIA’s team uses volumetric rendering to train the texture and geometry robustly, and second, it uses surface-based rendering to train only the texture for quality enhancement. Both stages use amortized optimization over prompts to maintain fast generation.

Nvidia’s Latte 3D generates text-to-3D in seconds!
Nvidia’s Latte 3D generates text-to-3D in seconds!

What sets Latte3D apart is its extensive pretraining phase, enabling the model to quickly adapt to new tasks by drawing on a vast repository of learned patterns and structures. This efficiency is achieved through a rigorous training regime that includes a blend of 3D datasets and prompts from ChatGPT.

Why does it matter?

AI models such as NVIDIA’s Latte3D have significantly reduced the time required to generate 3D visualizations from an hour to a few minutes compared to a few years ago. This technology has the potential to significantly accelerate the design and development process in various fields, such as the video game industry, advertising, and more.

Source

Quiet-STaR: LMs can self-train to think before responding

A groundbreaking study demonstrates the successful training of large language models (LM) to reason from text rather than specific reasoning tasks. The research introduces a novel training approach, Quiet STaR, which utilizes a parallel sampling algorithm to generate rationales from all token positions in a given string.

Quiet-STaR: LMs can self-train to think before responding
Quiet-STaR: LMs can self-train to think before responding

This technique integrates meta tokens to indicate when the LM should generate a rationale and when it should make a prediction based on the rationale, revolutionizing the understanding of LM behavior. Notably, the study shows that thinking enables the LM to predict difficult tokens more effectively, leading to improvements with longer thoughts.

The research introduces powerful advancements, such as a non-myopic loss approach, the application of a mixing head for retrospective determination, and the integration of meta tokens, underpinning a comprehensive leap forward in language model training.

Why does it matter?

These significant developments in language modeling advance the field and have the potential to revolutionize a wide range of applications. This points towards a future where large language models will unprecedentedly contribute to complex reasoning tasks.

Source

Neuralink’s first brain chip patient plays chess with his mind

Elon Musk’s brain chip startup, Neuralink, showcased its first brain chip patient playing chess using only his mind. The patient, Noland Arbaugh, was paralyzed below the shoulder after a diving accident.

Neuralink’s brain implant technology allows people with paralysis to control external devices using their thoughts. With further advancements, Neuralink’s technology has the potential to revolutionize the lives of people with paralysis, providing them with newfound independence and the ability to interact with the world in previously unimaginable ways.

Why does it matter?

Neuralink’s brain chip holds significant importance in AI and human cognition. It has the potential to enhance communication, assist paralyzed individuals, merge human intelligence with AI, and address the risks associated with AI development. However, ethical considerations and potential misuse of this technology must also be carefully examined.

Source

What Else Is Happening in AI on March 22nd, 2024❗

🤖 Microsoft introduces a new Copilot for better productivity.

Microsoft’s new Copilot for Windows and Surface devices is a powerful productivity tool integrating large language models with Microsoft Graph and Microsoft 365 apps to enhance work efficiency. With a focus on delivering AI responsibly while ensuring data security and privacy, Microsoft is dedicated to providing users with innovative tools to thrive in the evolving work landscape. (Link)

💰 Saudi Arabia to invest $40 billion in AI

Saudi Arabia has announced its plan to invest $40 billion in AI to become a global leader. Middle Eastern countries use their sovereign wealth fund, which has over $900 billion in assets, to achieve this goal. This investment aims to position the country at the forefront of the fast-evolving AI sector, drive innovation, and enhance economic growth. (Link)

🎧 Rightsify releases Hydra II to revolutionize AI music generation

Rightsify, a global music licensing leader, introduced Hydra II, the latest AI generation model. Hydra II offers over 800 instruments, 50 languages, and editing tools for customizable, copyright-free AI music. The model is trained on audio, text descriptions, MIDI, chord progressions, sheet music, and stems to create unique generations. (Link)

🚀 Open Interpreter’s 01 Light personal pocket AI agent

The Open Interpreter unveiled 01 Light, a portable device that allows you to control your computer using natural language commands. It’s part of an open-source project to make computing more accessible and flexible. It’s designed to make your online tasks more manageable, helping you get more done and simplify your life. (Link)

🤝 Microsoft’s $650 million Inflection deal: A strategic move
Microsoft has recently entered into a significant deal with AI startup Inflection, involving a payment of $650 million in cash. While the deal may seem like a licensing agreement, it appears to be a strategic move by Microsoft to acquire AI talent while avoiding potential regulatory trouble. (Link)

Microsoft unveiled its first “AI PCs,” with a dedicated Copilot key and Neural Processing Units (NPUs).

Microsoft unveiled its first "AI PCs," with a dedicated Copilot key and Neural Processing Units (NPUs).
Microsoft unveiled its first “AI PCs,” with a dedicated Copilot key and Neural Processing Units (NPUs).

Source: Nvidia

OpenAI Courts Hollywood in Meetings With Film Studios, Directors – from Bloomberg

The artificial intelligence startup has scheduled meetings in Los Angeles next week with Hollywood studios, media executives and talent agencies to form partnerships in the entertainment industry and encourage filmmakers to integrate its new AI video generator into their work, according to people familiar with the matter.

The upcoming meetings are just the latest round of outreach from OpenAI in recent weeks, said the people, who asked not to be named as the information is private. In late February, OpenAI scheduled introductory conversations in Hollywood led by Chief Operating Officer Brad Lightcap. Along with a couple of his colleagues, Lightcap demonstrated the capabilities of Sora, an unreleased new service that can generate realistic-looking videos up to about a minute in length based on text prompts from users. Days later, OpenAI Chief Executive Officer Sam Altman attended parties in Los Angeles during the weekend of the Academy Awards.

In an attempt to avoid defeatism, I’m hoping this will contribute to the indie boom with creatives refusing to work with AI and therefore studios who insist on using it. We’ve already got people on twitter saying this is the end of the industry but maybe only tentpole films as we know them.

Here’s the article without the paywall.

Catherine, the Princess of Wales, has cancer, she announced in a video message released by Kensington Palace on Friday March 22nd, 2024

The recent news surrounding Kate Middleton, the Princess of Wales, revolves around a manipulated family photo that sparked controversy and conspiracy theories. The photo, released by Middleton herself, depicted her with her three children and was met with speculation about potential AI involvement in its editing. However, experts suggest that the image was likely manipulated using traditional photo editing software like Photoshop rather than generative AI

The circumstances surrounding Middleton’s absence from the public eye due to abdominal surgery fueled rumors and intensified scrutiny over the edited photo.

Major news agencies withdrew the image, citing evidence of manipulation in areas like Princess Charlotte’s sleeve cuff and the alignment of elements in the photo.

Despite concerns over AI manipulation, this incident serves as a reminder that not all image alterations involve advanced technology, with this case being attributed to a botched Photoshop job.From an AI perspective, experts highlight how the incident reflects society’s growing awareness of AI technologies and their impact on shared reality. The controversy surrounding the edited photo underscores the need for transparency and accountability in media consumption to combat misinformation and maintain trust in visual content. As AI tools become more accessible and sophisticated, distinguishing between authentic and manipulated media becomes increasingly challenging, emphasizing the importance of educating consumers and technologists on identifying AI-generated content.Kate Middleton, the Princess of Wales, recently disclosed her battle with cancer in a heartfelt statement. Following major abdominal surgery in January, it was initially believed that her condition was non-cancerous. However, subsequent tests revealed the presence of cancer, leading to the recommendation for preventative chemotherapy. The 42-year-old princess expressed gratitude for the support received during this challenging time and emphasized the importance of privacy as she focuses on her treatment and recovery. The news of her diagnosis has garnered an outpouring of support from around the world, with messages of encouragement coming from various public figures and officials.

Nvidia CEO says we’ll see fully AI-generated games in 5-10 years

Nvidia’s CEO, Jensen Huang, predicts the emergence of fully AI-generated games within the next five to ten years. This prediction is based on the development of Nvidia’s next-generation Blackwell AI GPU, the B200. This GPU marks a significant shift in GPU usage towards creating neural networks for generating content rather than traditional rasterization or ray tracing for visual fidelity in games. The evolution of AI in gaming is highlighted as GPUs transition from rendering graphics to processing AI algorithms for content creation, indicating a major transformation in the gaming industry’s future landscape.The integration of AI into gaming represents a paradigm shift that could revolutionize game development and player experiences. Fully AI-generated games have the potential to offer unprecedented levels of customization, dynamic storytelling, and adaptive gameplay based on individual player interactions. This advancement hints at a new era of creativity and innovation in game design but also raises questions about the ethical implications and challenges surrounding AI-generated content, such as ensuring diversity, fairness, and avoiding biases in virtual worlds. Source

Andrew Ng, cofounder of Google Brain & former chief scientist @ Baidu- “I think AI agentic workflows will drive massive AI progress this year — perhaps even more than the next generation of foundation models.

This is an important trend, and I urge everyone who works in AI to pay attention to it.”

AI agentic workflows will drive massive AI progress
AI agentic workflows will drive massive AI progress

I think AI agentic workflows will drive massive AI progress this year — perhaps even more than the next generation of foundation models. This is an important trend, and I urge everyone who works in AI to pay attention to it.

Today, we mostly use LLMs in zero-shot mode, prompting a model to generate final output token by token without revising its work. This is akin to asking someone to compose an essay from start to finish, typing straight through with no backspacing allowed, and expecting a high-quality result. Despite the difficulty, LLMs do amazingly well at this task!

With an agentic workflow, however, we can ask the LLM to iterate over a document many times. For example, it might take a sequence of steps such as:

  • Plan an outline.

  • Decide what, if any, web searches are needed to gather more information.

  • Write a first draft.

  • Read over the first draft to spot unjustified arguments or extraneous information.

  • Revise the draft taking into account any weaknesses spotted.

  • And so on.

This iterative process is critical for most human writers to write good text. With AI, such an iterative workflow yields much better results than writing in a single pass.

Devin’s splashy demo recently received a lot of social media buzz. My team has been closely following the evolution of AI that writes code. We analyzed results from a number of research teams, focusing on an algorithm’s ability to do well on the widely used HumanEval coding benchmark. You can see our findings in the diagram below.

GPT-3.5 (zero shot) was 48.1% correct. GPT-4 (zero shot) does better at 67.0%. However, the improvement from GPT-3.5 to GPT-4 is dwarfed by incorporating an iterative agent workflow. Indeed, wrapped in an agent loop, GPT-3.5 achieves up to 95.1%.

Open source agent tools and the academic literature on agents are proliferating, making this an exciting time but also a confusing one. To help put this work into perspective, I’d like to share a framework for categorizing design patterns for building agents. My team AI Fund is successfully using these patterns in many applications, and I hope you find them useful.

  • Reflection: The LLM examines its own work to come up with ways to improve it.

  • Tool use: The LLM is given tools such as web search, code execution, or any other function to help it gather information, take action, or process data.

  • Planning: The LLM comes up with, and executes, a multistep plan to achieve a goal (for example, writing an outline for an essay, then doing online research, then writing a draft, and so on).

  • Multi-agent collaboration: More than one AI agent work together, splitting up tasks and discussing and debating ideas, to come up with better solutions than a single agent would.

  • Source

A daily chronicle of AI Innovations: March 21st, 2024 : 🕵️‍♂️ Stealing Part of a Production Language Model
🤖 Sakana AI’s method to automate foundation model development
👋 Key Stable Diffusion researchers leave Stability AI  🗣️Character AI’s new feature adds voice to characters with just 10-sec audio 💡Fitbit to get major AI upgrades powered by Google’s ‘Personal Health’ LLM 🔬Samsung creates lab to research chips for AI’s next phase 🤖GitHub’s latest AI tool can automatically fix code vulnerabilities

Stealing Part of a Production Language Model

Researchers from Google, OpenAI, and DeepMind (among others) released a new paper that introduces the first model-stealing attack that extracts precise, nontrivial information from black-box production language models like OpenAI’s ChatGPT or Google’s PaLM-2.

The attack allowed them to recover the complete embedding projection layer of a transformer language model. It differs from prior approaches that reconstruct a model in a bottom-up fashion, starting from the input layer. Instead, this operates top-down and directly extracts the model’s last layer by making targeted queries to a model’s API. This is useful for several reasons; it

  • Reveals the width of the transformer model, which is often correlated with its total parameter count.
  • Slightly reduces the degree to which the model is a complete “blackbox”
  • May reveal more global information about the model, such as relative size differences between different models

While there appear to be no immediate practical consequences of learning this layer is stolen, it represents the first time that any precise information about a deployed transformer model has been stolen.

Stealing Part of a Production Language Model
Stealing Part of a Production Language Model

Why does this matter?

Though it has limitations, the paper motivates the further study of practical attacks on ML models, in order to ultimately develop safer and more reliable AI systems. It also highlights how small, system-level design decisions impact the safety and security of the full product.

Source

Sakana AI’s method to automate foundation model development

Sakana AI has introduced Evolutionary Model Merge, a general method that uses evolutionary techniques to efficiently discover the best ways to combine different models from the vast ocean of different open-source models with diverse capabilities.

As of writing, Hugging Face has over 500k models in dozens of different modalities that, in principle, could be combined to form new models with new capabilities. By working with the vast collective intelligence of existing open models, this method is able to automatically create new foundation models with desired capabilities specified by the user.

Why does this matter?

Model merging shows great promise and democratizes up model-building. In fact, the current Open LLM Leaderboard is dominated by merged models. They work without any additional training, making it very cost-effective. But we need a more systematic approach.

Evolutionary algorithms, inspired by natural selection, can unlock more effective merging. They can explore vast possibilities, discovering novel and unintuitive combinations that traditional methods and human intuition might miss.

Source

Key Stable Diffusion researchers leave Stability AI

Robin Rombach and other key researchers who helped develop the Stable Diffusion text-to-image generation model have left the troubled, once-hot, now floundering GenAI startup.

Rombach (who led the team) and fellow researchers Andreas Blattmann and Dominik Lorenz were three of the five authors who developed the core Stable Diffusion research while at a German university. They were hired afterwards by Stability. Last month, they helped publish a 3rd edition of the Stable Diffusion model, which, for the first time, combined the diffusion structure used in earlier versions with transformers used in OpenAI’s ChatGPT.

Their departures are the latest in a mass exodus of executives at Stability AI, as its cash reserves dwindle and it struggles to raise additional funds.

Why does this matter?

Stable Diffusion is one of the foundational models that helped catalyze the boom in generative AI imagery, but now its future hangs in the balance. While Stability AI’s current situation raises questions about its long-term viability, the exodus potentially benefits its competitors.

Source

What Else Is Happening in AI on March 21st, 2024❗

🗣️Character AI’s new feature adds voice to characters with just 10-sec audio

You can now give voice to your Characters by choosing from thousands of voices or creating your own. The voices are created with just 10 seconds of audio clips. The feature is now available for free to everyone. (Link)

🤖GitHub’s latest AI tool can automatically fix code vulnerabilities

GitHub launches the first beta of its code-scanning autofix feature, which finds and fixes security vulnerabilities during the coding process. GitHub claims it can remediate more than two-thirds of the vulnerabilities it finds, often without the developers having to edit the code. The feature is now available for all GitHub Advanced Security (GHAS) customers. (Link)

GitHub’s latest AI tool can automatically fix code vulnerabilities
GitHub’s latest AI tool can automatically fix code vulnerabilities

🚀OpenAI plans to release a ‘materially better’ GPT-5 in mid-2024

According to anonymous sources from Businessinsider, OpenAI plans to release GPT-5 this summer, which will be significantly better than GPT-4. Some enterprise customers are said to have already received demos of the latest model and its ChatGPT improvements. (Link)

💡Fitbit to get major AI upgrades powered by Google’s ‘Personal Health’ LLM

Google Research and Fitbit announced they are working together to build a Personal Health LLM that gives users more insights and recommendations based on their data in the Fitbit mobile app. It will give Fitbit users personalized coaching and actionable insights that help them achieve their fitness and health goals. (Link)

🔬Samsung creates lab to research chips for AI’s next phase

Samsung has set up a research lab dedicated to designing an entirely new type of semiconductor needed for (AGI). The lab will initially focus on developing chips for LLMs with a focus on inference. It aims to release new “chip designs, an iterative model that will provide stronger performance and support for increasingly larger models at a fraction of the power and cost.” (Link)

A daily chronicle of AI Innovations: March 20th, 2024 : 🤖 OpenAI to release GPT-5 this summer; 🧠 Nvidia’s Jensen Huang says AI hallucinations are solvable, AGI is 5 years away; 🔬 Ozempic creator plans AI supercomputer to discover new drugs; 👀 After raising $1.3B, Inflection eaten alive by Microsoft; 🧠 MindEye2: AI Mind Reading from Brain Activity; 🚀 Nvidia NIM enables faster deployment of AI models

🤖 OpenAI to release GPT-5 this summer

  • OpenAI is planning to launch GPT-5 around mid-year, aiming to address previous performance issues and significantly improve upon its predecessor, GPT-4.
  • GPT-5 is described as “materially better” by those who have seen demos, including enhancements and new capabilities like the ability to call AI agents for autonomous tasks, with enterprise customers having already previewed these improvements.
  • The release timeline for GPT-5 remains uncertain as OpenAI continues its training and thorough safety and vulnerability testing, with no specific deadline for completion of these preparatory steps.
  • Source

👀 After raising $1.3B, Inflection eaten alive by Microsoft 

  • In June 2023, Inflection raised $1.3 billion led by Microsoft to develop “more personal AI” but was overtaken by Microsoft less than a year later, with co-founders joining Microsoft’s new AI division.
  • Despite significant investment, Inflection’s AI, Pi, failed to compete with advancements from other companies such as OpenAI, Google’s Gemini, and Anthropic, leading to its downfall.
  • Microsoft’s takeover of Inflection reflects the strategy of legacy tech companies to dominate the AI space by supporting startups then acquiring them once they face challenges.
  • Source

🧠 Nvidia’s Jensen Huang says AI hallucinations are solvable, AGI is 5 years away

  • Nvidia CEO Jensen Huang predicts artificial general intelligence (AGI) could be achieved within 5 years, depending on how AGI is defined and measured.
  • Huang addresses concerns around AI hallucinations, suggesting that ensuring answers are well-researched could easily solve the issue.
  • The concept of AGI raises concerns about its potential unpredictability and the challenges of aligning its objectives with human values and priorities.
  • Source

🔬 Ozempic creator plans AI supercomputer to discover new drugs

  • The Novo Nordisk Foundation is investing in “Gefion,” an AI supercomputer project developed in collaboration with Nvidia.
  • “Gefion” aims to be the world’s most powerful AI supercomputer for health sciences, utilizing Nvidia’s new chips to accelerate scientific breakthroughs in critical areas such as drug discovery, disease diagnosis, and treatment,
  • This initiative underscores the growing integration of AI in healthcare, promising to catalyze significant scientific discoveries and innovations that could transform patient care and outcomes.
  • Source

MindEye2: AI mind reading from brain activity

MindEye2 is a revolutionary model that reconstructs visual perception from brain activity using just one hour of data. Traditional methods require extensive training data, making them impractical for real-world applications. However, MindEye2 overcomes this limitation by leveraging shared-subject models. The model is pretrained on data from seven subjects and then fine-tuned with minimal data from a new subject.

MindEye2: AI mind reading from brain activity
MindEye2: AI mind reading from brain activity

By mapping brain activity to a shared-subject latent space and then nonlinear mapping to CLIP image space, MindEye2 achieves high-quality reconstructions with limited training data. It performs state-of-the-art image retrieval and reconstruction across multiple subjects within only 2.5% of the previously required training data, reducing the training time from 40 to just one hour.

Why does it matter?

MindEye2 has the potential to revolutionize clinical assessments and brain-computer interface applications. This remarkable achievement also holds great promise for neuroscience and opens new possibilities for understanding how our brains perceive and process visual information. It can also help develop personalized treatment plans for neuro patients.

Source

Nvidia NIM enables faster deployment of AI models 

NVIDIA has introduced NVIDIA NIM (NVIDIA Inference Microservices) to accelerate the deployment of AI applications for businesses. NIM is a collection of microservices that package essential components of an AI application, including AI models, APIs, and libraries, into a container. These containers can be deployed in environments such as cloud platforms, Linux servers, or serverless architectures.

Nvidia NIM enables faster deployment of AI models 
Nvidia NIM enables faster deployment of AI models

NIM significantly reduces the time it takes to deploy AI applications from weeks to minutes. It offers optimized inference engines, industry-standard APIs, and support for popular software and data platform vendors. NIM microservices are compatible with NVIDIA GPUs and support features like Retrieval Augmented Generation (RAG) capabilities for enhanced enterprise applications. Developers can experiment with NIM microservices for free on the ai.nvidia.com platform, while commercial deployment is available through NVIDIA AI Enterprise 5.0.

Why does it matter?

With NIM, Nvidia is trying to democratize AI deployment for enterprises by abstracting away complexities. This will enable more developers to contribute to their company’s AI transformation efforts and allow businesses to run AI applications almost instantly without specialized AI expertise.

Source

Microsoft hires DeepMind co-founder to lead a new AI division

Mustafa Suleyman, a renowned co-founder of DeepMind and Inflection, has recently joined Microsoft as the leader of Copilot. Satya Nadella, Microsoft’s CEO, made this significant announcement, highlighting the importance of innovation in artificial intelligence (AI).

In his new role as the Executive Vice President and CEO of Microsoft AI, Mustafa will work alongside Karén Simonyan, another talented individual from Inflection who will serve as Chief Scientist. Together, they will spearhead the development and advancement of Copilot and other exciting consumer AI products at Microsoft. Mustafa and his team’s addition to the Microsoft family brings a wealth of expertise and promises groundbreaking advancements in AI.

Why does it matter?

Mustafa Suleyman’s expertise in AI is expected to contribute to the development of innovative consumer AI products and research at Microsoft, furthering its mission to bring the benefits of AI to people and organizations worldwide. With DeepMind’s founder now at the helm, the AI race between Microsoft, Google, and others became even more intense.

Source

What Else Is Happening in AI on March 20th, 2024❗

📞 Truecaller adds AI-powered spam detection and blocking for Android users

Truecaller has unveiled a new feature for its Android premium subscribers that uses AI to detect spam, even if unavailable on the Truecaller database, and block every call that doesn’t come from an approved contact. Truecaller hopes to add more premium subscribers to its list by adding this feature. However, this feature is not available for Apple users. (Link)

⚽ Google DeepMind’s new AI tool can analyze soccer tactics and offer insights 

DeepMind has partnered with Liverpool FC to develop a new AI tool called TacticAI. TacticAI uses generative and predictive AI to help coaches determine which player will most likely receive the ball during corner kicks, whether a shot will be taken, and how to adjust player setup. It aims to revolutionize soccer and help the teams enhance their efficiency. (Link)

🎬 Pika Labs introduces sound effects for its gen-AI video generation

Pika Labs has now added the ability to create sound effects from a text prompt for its generative artificial intelligence videos. It allows for automatic or custom SFX generations to pair with video outputs. Now, users can make bacon sizzle, lions roar, or add footsteps to the video of someone walking down the street. It is only available to pro users. (Link)

🎮 Buildbox 4 Alpha enables users to create 3D video games from text prompts 

Buildbox has released an alpha version of Buildbox 4. It’s an AI-first game engine that allows users to create games and generate assets from text prompts. The alpha version aims to make text-to-game a distinct reality. Users can create various assets and animations from simple text prompts. It also allows users to build a gaming environment in a few minutes. (Link)

🤖 Nvidia adds generative AI capabilities to empower humanoid robots

Nvidia introduced Project GR00T, a multimodal AI that will power future humanoids with advanced foundation AI. Project GR00T enables humanoid robots to input text, speech, videos, or even live demos and process them to take specific actions. It has been developed with the help of Nvidia’s Isaac Robotic Platform tools, including an Isaac Lab for RLHF. (Link)

The EU AI Act – Key takeaways for LLM builders

The EU AI Act - Key takeaways for LLM builders

A daily chronicle of AI Innovations: March 19th, 2024 : 💻 Nvidia launches ‘world’s most powerful AI chip’; 🎥 Stability AI’s SV3D turns a single photo into a 3D video; 🤖 OpenAI CEO hints at “Amazing Model”, maybe ChatGPT-5 ;🤝 Apple is in talks to bring Google’s AI to iPhones

Nvidia launches ‘world’s most powerful AI chip’

Nvidia has revealed its new Blackwell B200 GPU and GB200 “superchip”, claiming it to be the world’s most powerful chip for AI. Both B200 and GB200 are designed to offer powerful performance and significant efficiency gains.

Nvidia launches 'world's most powerful AI chip'
Nvidia launches ‘world’s most powerful AI chip’

Key takeaways:

  • The B200 offers up to 20 petaflops of FP4 horsepower, and Nvidia says it can reduce costs and energy consumption by up to 25 times over an H100.
  • The GB200 “superchip” can deliver 30X the performance for LLM inference workloads while also being more efficient.
  • Nvidia claims that just 2,000 Blackwell chips working together could train a GPT -4-like model comprising 1.8 trillion parameters in just 90 days.

Why does this matter?

A major leap in AI hardware, the Blackwell GPU boasts redefined performance and energy efficiency. This could lead to lower operating costs in the long run, making high-performance computing more accessible for AI research and development, all while promoting eco-friendly practices.

Source

Stability AI’s SV3D turns a single photo into a 3D video

Stability AI released Stable Video 3D (SV3D), a new generative AI tool for rendering 3D videos. SV3D can create multi-view 3D models from a single image, allowing users to see an object from any angle. This technology is expected to be valuable in the gaming sector for creating 3D assets and in e-commerce for generating 360-degree product views.

SV3D builds upon Stability AI’s previous Stable Video Diffusion model. Unlike prior methods, SV3D can generate consistent views from any given angle. It also optimizes 3D meshes directly from the novel views it produces.

SV3D comes in two variants: SV3D_u generates orbital videos from single images, and SV3D_p creates 3D videos along specified camera paths.

Why does this matter?

SV3D represents a significant leap in generative AI for 3D content. Its ability to create 3D models and videos from a single image could open up possibilities in various fields, such as animation, virtual reality, and scientific modeling.

Source

OpenAI CEO hints at “Amazing Model,” maybe ChatGPT-5

OpenAI CEO Sam Altman has announced that the company will release an “amazing model” in 2024, although the name has not been finalized. Altman also mentioned that OpenAI plans to release several other important projects before discussing GPT-5, one of which could be the Sora video model.

Source

Altman declined to comment on the Q* project, which is rumored to be an AI breakthrough related to logic. He also expressed his opinion that GPT-4 Turbo and GPT-4 “kind of suck” and that the jump from GPT-4 to GPT-5 could be as significant as the improvement from GPT-3 to GPT-4.

Why does this matter?

This could mean that after Google Gemini and Claude-3’s latest version, a new model, possibly ChatGPT-5, could be released in 2024. Altman’s candid remarks about the current state of AI models also offer valuable context for understanding the anticipated advancements and challenges in the field.

Source

Project GR00T is an ambitious initiative aiming to develop a general-purpose foundation model for humanoid robot learning, addressing embodied AGI challenges. Collaborating with leading humanoid companies worldwide, GR00T aims to understand multimodal instructions and perform various tasks.

GROOT is a foundational model that takes language, videos, and example demonstrations as inputs so it can produce the next action.

What the heck does that mean?

➡️ It means you can show it how to do X a few times, and then it can do X on its own.

Like cooking, drumming, or…

Source

Google’s new fine-tuned model is a HUGE improvement, AI is coming for human doctors sooner than most believe.

Google's new fine-tuned model is a HUGE improvement, AI is coming for human doctors sooner than most believe.
Google’s new fine-tuned model is a HUGE improvement, AI is coming for human doctors sooner than most believe.

NVIDIA creates Earth-2 digital twin: generative AI to simulate, visualize weather and climate. Source

What Else Is Happening in AI on March 19th, 2024❗

🤝 Apple is in talks to bring Google’s AI to iPhones

Apple and Google are negotiating a deal to integrate Google’s Gemini AI into iPhones, potentially shaking up the AI industry. The deal would expand on their existing search partnership. Apple also held discussions with OpenAI. If successful, the partnership could give Gemini a significant edge with billions of potential users. (Link)

 🏷️YouTube rolls out AI content labels

YouTube now requires creators to self-label AI-generated or synthetic content in videos. The platform may add labels itself for potentially misleading content. However, the tool relies on creators being honest, as YouTube is still working on AI detection tools. (Link)

🎮Roblox speeds up 3D creation with AI tools
Roblox has introduced two AI-driven tools to streamline 3D content creation on its platform. Avatar Auto Setup automates the conversion of 3D body meshes into fully animated avatars, while Texture Generator allows creators to quickly alter the appearance of 3D objects using text prompts, enabling rapid prototyping and iteration.(Link)

🌐Nvidia teams up with Shutterstock and Getty Images for AI-generated 3D content

Nvidia’s Edify AI can now create 3D content, and partnerships with Shutterstock and Getty Images will make it accessible to all. Developers can soon experiment with these models, while industry giants are already using them to create stunning visuals and experiences.  (Link)

🖌️Adobe Substance 3D introduces AI-powered text-to-texture tools

Adobe has introduced two AI-driven features to its Substance 3D suite: “Text to Texture,” which generates photo-realistic or stylized textures from text prompts, and “Generative Background,” which creates background images for 3D scenes. Both tools use 2D imaging technology from Adobe’s Firefly AI model to streamline 3D workflows. (Link)

A daily chronicle of AI Innovations: March 18th, 2024 – Bernie’s 4 day workweek: less work, same pay – Google’s AI brings photos to life as talking avatars – Elon Musk’s xAI open-sources Grok AI

Bernie’s 4 day workweek: less work, same pay

Sen. Bernie Sanders has introduced the Thirty-Two Hour Workweek Act, which aims to establish a four-day workweek in the United States without reducing pay or benefits. To be phased in over four years, the bill would lower the overtime pay threshold from 40 to 32 hours, ensuring that workers receive 1.5 times their regular salary for work days longer than 8 hours and double their regular wage for work days longer than 12 hours.

Sanders, along with Sen. Laphonza Butler and Rep. Mark Takano, believes that this bill is crucial in ensuring that workers benefit from the massive increase in productivity driven by AI, automation, and new technology. The legislation aims to reduce stress levels and improve Americans’ quality of life while also protecting their wages and benefits.

Why does this matter?

This bill could alter the workforce dynamics. Businesses may need to assess staffing and invest in AI to maintain productivity. While AI may raise concerns over job displacements, it also offers opportunities for better work-life balance through efficiency gains by augmenting human capabilities.

Source

Google’s AI brings photos to life as talking avatars

Google’s latest AI research project VLOGGER, automatically generates realistic videos of talking and moving people from just a single image and an audio or text input. It is the first model that aims to create more natural interactions with virtual agents by including facial expressions, body movements, and gestures, going beyond simple lip-syncing.

It uses a two-step process: first, a diffusion-based network predicts body motion and facial expressions based on the audio, and then a novel architecture based on image diffusion models generates the final video while maintaining temporal consistency. VLOGGER outperforms previous state-of-the-art methods in terms of image quality, diversity, and the range of scenarios it can handle.

Why does this matter?

VLOGGER’s flexibility and applications could benefit remote work, education, and social interaction, making them more inclusive and accessible. Also, as AR/VR technologies advance, VLOGGER’s avatars could create emotionally resonant experiences in gaming, entertainment, and professional training scenarios.

Source

Elon Musk’s xAI open-sources Grok AI

Elon Musk’s xAI has open-sourced the base model weights and architecture of its AI chatbot, Grok. This allows researchers and developers to freely use and build upon the 314 billion parameter Mixture-of-Experts model. Released under the Apache 2.0 license, the open-source version is not fine-tuned for any particular task.

Why does this matter?

This move aligns with Musk’s criticism of companies that don’t open-source their AI models, including OpenAI, which he is currently suing for allegedly breaching an agreement to remain open-source. While several fully open-source AI models are available, the most used ones are closed-source or offer limited open licenses.

Source

What Else Is Happening in AI on March 18th, 2024❗

🧠 Maisa KPU may be the next leap in AI reasoning

Maisa has released the beta version of its Knowledge Processing Unit (KPU), an AI system that uses LLMs’ advanced reasoning and data processing abilities. In an impressive demo, the KPU assisted a customer with an order-related issue, even when the customer provided an incorrect order ID, showing the system’s understanding abilities. (Link)

🍿 PepsiCo increases market domination using GenAI

PepsiCo uses GenAI in product development and marketing for faster launches and better profitability. It has increased market penetration by 15% by using GenAI to improve the taste and shape of products like Cheetos based on customer feedback. The company is also doubling down on its presence in India, with plans to open a third capability center to develop local talent. (Link)

💻 Deci launches Nano LLM & GenAI dev platform

Israeli AI startup Deci has launched two major offerings: Deci-Nano, a small closed-source language model, and a complete Generative AI Development Platform for enterprises. Compared to rivals like OpenAI and Anthropic, Deci-Nano offers impressive performance at low cost, and the new platform offers a suite of tools to help businesses deploy and manage AI solutions. (Link)

🎮 Invoke AI simplifies game dev workflows

Invoke has launched Workflows, a set of AI tools designed for game developers and large studios. These tools make it easier for teams to adopt AI, regardless of their technical expertise levels. Workflows allow artists to use AI features while maintaining control over their training assets, brand-specific styles, and image security. (Link)

🚗 Mercedes teams up with Apptronik for robot workers

Mercedes-Benz is collaborating with robotics company Apptronik to automate repetitive and physically demanding tasks in its manufacturing process. The automaker is currently testing Apptronik’s Apollo robot, a 160-pound bipedal machine capable of lifting objects up to 55 pounds. The robot inspects and delivers components to human workers on the production line, reducing the physical strain on employees and increasing efficiency. (Link)

A daily chronicle of AI Innovations: Week 2 Recap

  1. DeepSeek released DeepSeek-VL, an open-source Vision-Language (VL) model designed for real-world vision and language understanding applications. The DeepSeek-VL family, includes 7B and1.3B base and chat models and achieves state-of-the-art or competitive performance across a wide range of visual-language benchmarks. Free for commercial use [Details | Hugging Face | Demo]

  2. Cohere released Command-R, a 35 billion parameters generative model with open weights, optimized for long context tasks such as retrieval augmented generation (RAG) and using external APIs and tools for production-scale AI for enterprise [Details | Hugging Face].

  3. Google DeepMind introduced SIMA (Scalable Instructable Multiworld Agent), a generalist AI agent for 3D virtual environments, trained on nine different video games. It can understand a broad range of gaming worlds, and follows natural-language instructions to carry out tasks within them, as a human might.  It doesn’t need access to a game’s source code or APIs and requires only the images on screen, and natural-language instructions provided by the user. SIMA uses keyboard and mouse outputs to control the games’ central character to carry out these instructions [Details].

  4. Meta AI introduces Emu Video Edit (EVE), a model that establishes a new state-of-the art in video editing without relying on any supervised video editing data [Details].

  5. Cognition Labs introduced Devin, the first fully autonomous AI software engineer. Devin can learn how to use unfamiliar technologies, can build and deploy apps end to end, can train and fine tune its own AI models. When evaluated on the SWE-Bench benchmark, which asks an AI to resolve GitHub issues found in real-world open-source projects, Devin correctly resolves 13.86% of the issues unassisted, exceeding the previous state-of-the-art model performance of 1.96% unassisted and 4.80% assisted [Details].

  6. Pika Labs adds sound effects to its AI video tool, Pika, allowing users to either prompt desired sounds or automatically generate them based on video content. [Video link].

  7. Anthropic’s Claude 3 Opus ranks #1 on LMSYS Chatbot Arena Leaderboard, along with GPT-4 [Link].

  8. The European Parliament approved the Artificial Intelligence Act. The new rules ban certain AI applications including biometric categorisation systems, Emotion recognition in the workplace and schools, social scoring and more [Details].

  9. Huawei Noah’s Ark Lab introduced PixArt–Σ, a Diffusion Transformer model (DiT) capable of directly generating images at 4K resolution. It achieves superior image quality and user prompt adherence with significantly smaller model size (0.6B parameters) than existing text-to-image diffusion models, such as SDXL (2.6B parameters) and SD Cascade (5.1B parameters) [Details].

  10. South Korean startup Hyodol AI has launched a $1,800 LLM-powered companion doll specifically designed to offer emotional support and companionship to the rapidly expanding elderly demographic in the country [Details].

  11. Covariant introduced RFM-1 (Robotics Foundation Model -1), a large language model (LLM), but for robot language. Set up as a multimodal any-to-any sequence model, RFM-1 is an 8 billion parameter transformer trained on text, images, videos, robot actions, and a range of numerical sensor readings [Details].

  12. Figure 01 robot integrated with an OpenAI vision-language model can now have full conversations with people [Link]

  13. Deepgram announced the general availability of Aura, a text-to-speech model built for responsive, conversational AI agents and applications [Details | Demo].

  14. Claude 3 Haiku model is now available alongside Sonnet and Opus in the Claude API and on claude.ai for Pro subscribers. Haiku outperforms GPT-3.5 and Gemini 1.0 pro while costing less, and is three times faster than its peers for the vast majority of workloads [Details].

  15. Paddle announced AI Launchpad, a 6-week remote program for AI founders to launch and scale an AI business with $20,000 in cash prize [Details].

  16. Midjourney adds feature for generating consistent characters across multiple gen AI images [Details].

  17. The Special Committee of the OpenAI Board announced the completion of the review. Altman, Brockman to continue to lead OpenAI [Details]

  18. Together.ai introduced Sequoia, a scalable, robust, and hardware-aware speculative decoding framework that improves LLM inference speed on consumer GPUs (with offloading), as well as on high-end GPUs (on-chip), without any approximations [Details].

  19. OpenAI released Transformer Debugger (TDB), a tool developed and used internally by OpenAI’s Superalignment team for investigating into specific behaviors of small language models [GitHub].

  20. Elon Musk announced that xAI will open source Grok this week [Link].

A Daily Chronicle of AI Innovations – March 16th, 2024:

🔍 FTC is probing Reddit’s AI licensing deals

  • Reddit is under investigation by the FTC for its data licensing practices concerning user-generated content being used to train AI models.
  • The investigation focuses on Reddit’s engagement in selling, licensing, or sharing data with third parties for AI training.
  • Reddit anticipates generating approximately USD 60 million in 2024 from a data licensing agreement with Google, aiming to leverage its platform data for training LLMs

.

💻 New jailbreak uses ASCII art to elicit harmful responses from leading LLM

  • Researchers identified a new vulnerability in leading AI language models, named ArtPrompt, which uses ASCII art to exploit the models’ security mechanisms.
  • ArtPrompt masks security-sensitive words with ASCII art, fooling language models like GPT-3.5, GPT-4, Gemini, Claude, and Llama2 into performing actions they would otherwise block, such as giving instructions for making a bomb.
  • The study underscores the need for enhanced defensive measures for language models, as ArtPrompt, by leveraging a mix of text-based and image-based inputs, can effectively bypass current security protocols.

OpenAI aims to make its own AI processors — chip venture in talks with Abu Dhabi investment firm. Source

Once “too scary” to release, GPT-2 gets squeezed into an Excel spreadsheet. Source

 A Daily Chronicle of AI Innovations – March 15th, 2024:

🍎 Apple quietly acquires another AI startup

🤖 Mercedes tests humanoid robots for ‘low skill, repetitive’ tasks

🚫 Midjourney bans prompts with Joe Biden and Donald Trump over election misinformation concerns

💰 El Salvador stashes $406 million in bitcoin in ‘cold wallet’

🤔 Microsoft calls out Google dominance in generative AI

📝 Anthropic releases affordable, high-speed Claude 3 Haiku model

🥘 Apple’s MM1: The new recipe to master AI performance

Apple’s MM1 AI model shows state-of-the-art language and vision capabilities. It was trained on a filtered dataset of 500 million text-image pairs from the web, including 10% text-only docs to improve language understanding.

🥘 Apple’s MM1: The new recipe to master AI performance
🥘 Apple’s MM1: The new recipe to master AI performance

The team experimented with different configurations during training. They discovered that using an external pre-trained high-resolution image encoder improved visual recognition. Combining different image, text, and caption data ratios led to the best performance. Synthetic caption data also enhanced few-shot learning abilities.

This experiment cements that using a blend of image caption, interleaved image text, and text-only data is crucial for achieving state-of-the-art (SOTA) few-shot results across multiple benchmarks.

Why does it matter?

Apple’s new model is promising, especially in developing image recognition systems for new categories or domains. This will help businesses and startups improve the speed of AI tool development specifically for text-to-image, document analysis, and enhanced visual recognition.

⚡ Cerebras WSE-3: AI chip enabling 10x larger models than GPT-4

Cerebras Systems has made a groundbreaking announcement unveiling its latest wafer-scale AI chip, the WSE-3. This chip boasts an incredible 4 trillion transistors, making it one of the most powerful AI chips on the market. The third-generation wafer-scale AI mega chip is twice as powerful as its predecessor while being power efficient. 

The chip’s transistor density has increased by over 50 percent thanks to the latest manufacturing technology. One of the most remarkable features of the WSE-3 chip is its ability to enable AI models that are ten times larger than the highly acclaimed GPT-4 and Gemini models.

Why does it matter?

The WSE-3 chip opens up new possibilities for tackling complex problems and pushing the boundaries of AI capabilities. This powerful system can train massive language models, such as the Llama 70B, in just one day. It will help enterprises create custom LLMs, rapidly reducing the time-to-market.

🤖 Apple acquires Canadian AI startup DarwinAI

Apple made a significant acquisition earlier this year by purchasing Canadian AI startup DarwinAI. Integrating DarwinAI’s expertise and technology bolsters Apple’s AI initiatives. 

With this acquisition, Apple aims to tap into DarwinAI’s advancements in AI technology, particularly in visual inspection during manufacturing and making AI systems smaller and faster. Leveraging DarwinAI’s technology, Apple aims to run AI on devices rather than relying solely on cloud-based solutions.

Why does it matter?

Apple’s acquisition of DarwinAI is a strategic move to revolutionize features and enhance its AI capabilities across various products and services. Especially with the iOS 18 release around the corner, this acquisition will help create new features and enhance the user experience.

🤖 Microsoft expands the availability of Copilot across life and work.

Microsoft is expanding Copilot, its AI assistant, with the introduction of the Copilot Pro subscription for individuals, the availability of Copilot for Microsoft 365 to small and medium-sized businesses, and removing seat minimums for commercial plans. Copilot aims to enhance creativity, productivity, and skills across work and personal life, providing users access to the latest AI models and improved image creation

💻 Oracle adds groundbreaking Generative AI features to its software

Oracle has added advanced AI capabilities to its finance and supply chain software suite, aimed at improving decision-making and enhancing customer and employee experience. For instance, Oracle Fusion Cloud SCM includes features such as item description generation, supplier recommendations, and negotiation summaries.

💰 Databricks makes a strategic investment in Mistral AI

Databricks has invested in Mistral AI and integrated its AI models into its data intelligence platform, allowing users to customize and consume models in various ways. The integration includes Mistral’s text-generation models, such as Mistral 7B and Mixtral 8x7B, which support multiple languages. This partnership aims to provide Databricks customers with advanced capabilities to leverage AI models and drive innovation in their data-driven applications.

📱 Qualcomm emerges as a mobile AI juggernaut

Qualcomm has solidified its leadership position in mobile artificial intelligence (AI). It has been developing AI hardware and software for over a decade. Their Snapdragon processors are equipped with specialized AI engines like Hexagon DSP, ensuring efficient AI and machine learning processing without needing to send data to the cloud.

👓 MIT researchers develop peripheral vision capabilities for AI models

AI researchers are developing techniques to simulate peripheral vision and improve object detection in the periphery. They created a new dataset to train computer vision models, which led to better object detection outside the direct line of sight, though still behind human capabilities. A modified texture tiling approach accurately representing information loss in peripheral vision significantly enhanced object detection and recognition abilities.

🤔 Microsoft calls out Google dominance in generative AI 

  • Microsoft has expressed concerns to EU antitrust regulators about Google’s dominance in generative AI, highlighting Google’s unique position due to its vast data sets and vertical integration, which includes AI chips and platforms like YouTube.
  • The company argues that Google’s control over vast resources and its own AI developments give it a competitive advantage, making it difficult for competitors to match, especially in the development of Large Language Models like Gemini.
  • Microsoft defends partnerships with startups like OpenAI as essential for innovation and competition in the AI market, countering regulatory concerns about potential anticompetitive advantages arising from such collaborations.

🤖 Mercedes tests humanoid robots for ‘low skill, repetitive’ tasks

  • Mercedes-Benz is testing humanoid robots, specifically Apptronik’s bipedal robot Apollo, for automating manual labor tasks in manufacturing.
  • The trial aims to explore the use of Apollo in physically demanding, repetitive tasks within existing manufacturing facilities without the need for significant redesigns.
  • The initiative seeks to address labor shortages by using robots for low-skill tasks, allowing highly skilled workers to focus on more complex aspects of car production.

🚫 Midjourney bans prompts with Joe Biden and Donald Trump over election misinformation concerns

  • Midjourney, an AI image generator, has banned prompts containing the names of Joe Biden and Donald Trump to avoid the spread of election misinformation.
  • The policy change is in response to concerns over AI’s potential to influence voters and spread false information before the 2024 presidential election.
  • Despite the new ban, Midjourney previously allowed prompts that could generate misleading or harmful content, and it was noted for its poor performance in controlling election disinformation.

Midjourney introduces Character Consistency: Tutorial

midjourney_character_consistency

A Daily Chronicle of AI Innovations – March 14th, 2024: 

🎮 DeepMind’s SIMA: The AI agent that’s a Jack of all games

 ⚡ Claude 3 Haiku: Anthropic’s lightning-fast AI solution for enterprises

 🤖 OpenAI-powered “Figure 01” can chat, perceive, and complete tasks

 🎥 OpenAI’s Sora will be publicly available later this year

 

🎮 DeepMind’s SIMA: The AI agent that’s a Jack of all games

DeepMind has introduced SIMA (Scalable Instructable Multiworld Agent), a generalist AI agent that can understand and follow natural language instructions to complete tasks across video game environments. Trained in collaboration with eight game studios on nine different games, SIMA marks a significant milestone in game-playing AI by showing the ability to generalize learned skills to new gaming worlds without requiring access to game code or APIs.

 

DeepMind's SIMA: The AI agent that's a Jack of all games
DeepMind’s SIMA: The AI agent that’s a Jack of all games
 

(SIMA comprises pre-trained vision models, and a main model that includes a memory and outputs keyboard and mouse actions.)

SIMA was evaluated on 600 basic skills, including navigation, object interaction, and menu use. In tests, SIMA agents trained on multiple games significantly outperformed specialized agents trained on individual games. Notably, an agent trained on all but one game performed nearly as well on the unseen game as an agent specifically trained on it, showcasing SIMA’s remarkable ability to generalize to new environments. 

Why does this matter?

SIMA’s generalization ability using a single AI agent is a significant milestone in transfer learning. By showing that a multi-task trained agent can perform nearly as well on an unseen task as a specialized agent, SIMA paves the way for more versatile and scalable AI systems. This could lead to faster deployment of AI in real-world applications, as agents would require less task-specific training data and could adapt to new scenarios more quickly.

Source


⚡ Claude 3 Haiku: Anthropic’s lightning-fast AI solution for enterprises

Anthropic has released Claude 3 Haiku, their fastest and most affordable AI model. With impressive vision capabilities and strong performance on industry benchmarks, Haiku is designed to tackle a wide range of enterprise applications. The model’s speed – processing 21K tokens per second for prompts under 32K tokens – and cost-effective pricing model make it an attractive choice for businesses needing to analyze large datasets and generate timely outputs.

 

Claude 3 Haiku: Anthropic's lightning-fast AI solution for enterprises
Claude 3 Haiku: Anthropic’s lightning-fast AI solution for enterprises
 

In addition to its speed and affordability, Claude 3 Haiku prioritizes enterprise-grade security and robustness. The model is now available through Anthropic’s API or on claude.ai for Claude Pro subscribers.

Why does this matter?

Claude 3 Haiku sets a new benchmark for enterprise AI by offering high speed and cost-efficiency without compromising performance. This release will likely intensify competition among AI providers, making advanced AI solutions more accessible to businesses of all sizes. As more companies adopt models like Haiku, we expect a surge in AI-driven productivity and decision-making across industries.

Source


🤖 OpenAI-powered “Figure 01” can chat, perceive, and complete tasks

Robotics company Figure, in collaboration with OpenAI, has developed a groundbreaking robot called “Figure 01” that can engage in full conversations, perceive its surroundings, plan actions, and execute tasks based on verbal requests, even those that are ambiguous or context-dependent. This is made possible by connecting the robot to a multimodal AI model trained by OpenAI, which integrates language and vision.

OpenAI-powered "Figure 01" can chat, perceive, and complete tasks
OpenAI-powered “Figure 01” can chat, perceive, and complete tasks

The AI model processes the robot’s entire conversation history, including images, enabling it to generate appropriate verbal responses and select the most suitable learned behaviors to carry out given commands. The robot’s actions are controlled by visuomotor transformers that convert visual input into precise physical movements. “Figure 01” successfully integrates natural language interaction, visual perception, reasoning, and dexterous manipulation in a single robot platform.

Why does this matter?

As robots become more adept at understanding and responding to human language, questions arise about their autonomy and potential impact on humanity. Collaboration between the robotics industry and AI policymakers is needed to establish regulations for the safe deployment of AI-powered robots. If deployed safely, these robots could become trusted partners, enhancing productivity, safety, and quality of life in various domains.

Source

What Else Is Happening in AI on March 14th, 2024❗

🛍️ Amazon streamlines product listing process with new AI tool

Amazon is introducing a new AI feature for sellers to quickly create product pages by pasting a link from their external website. The AI generates product descriptions and images based on the linked site’s information, saving sellers time. (Link)

🛡️ Microsoft to expand AI-powered cybersecurity tool availability from April 1

Microsoft is expanding the availability of its AI-powered cybersecurity tool, “Security Copilot,” from April 1, 2024. The tool helps with tasks like summarizing incidents, analyzing vulnerabilities, and sharing information. Microsoft plans to adopt a ‘pay-as-you-go’ pricing model to reduce entry barriers. (Link)

🎥 OpenAI’s Sora will be publicly available later this year

OpenAI will release Sora, its text-to-video AI tool, to the public later this year. Sora generates realistic video scenes from text prompts and may add audio capabilities in the future. OpenAI plans to offer Sora at a cost similar to DALL-E, its text-to-image model, and is developing features for users to edit the AI-generated content. (Link)

📰 OpenAI partners with Le Monde, Prisa Media for news content in ChatGPT

OpenAI has announced partnerships with French newspaper Le Monde and Spanish media group Prisa Media to provide their news content to users of ChatGPT. The media companies see this as a way to ensure reliable information reaches AI users while safeguarding their journalistic integrity and revenue. (Link)

🏠 Icon’s AI architect and 3D printing breakthroughs reimagine homebuilding

Construction tech startup Icon has introduced an AI-powered architect, Vitruvius, that engages users in designing their dream homes, offering 3D-printed and conventional options. The company also debuted an advanced 3D printing robot called Phoenix and a low-carbon concrete mix as part of its mission to make homebuilding more affordable, efficient, and sustainable. (Link)

A Daily Chronicle of AI Innovations – March 13th, 2024: Devin: The first AI software engineer redefines coding; Deepgram’s Aura empowers AI agents with authentic voices; Meta introduces two 24K GPU clusters to train Llama 3

Devin: The first AI software engineer redefines coding 

In the most groundbreaking development, the US-based startup Cognition AI has unveiled Devin, the world’s first AI software engineer. It is an autonomous agent that solves engineering tasks using its shell or command prompt, code editor, and web browser. Devin can also perform tasks like planning, coding, debugging, and deploying projects autonomously.

https://twitter.com/i/status/1767548763134964000

When evaluated on the SWE-Bench benchmark, which asks an AI to resolve GitHub issues found in real-world open-source projects, Devin correctly resolves 13.86% of the issues unassisted, far exceeding the previous state-of-the-art model performance of 1.96% unassisted and 4.80% assisted. It has successfully passed practical engineering interviews with leading AI companies and even completed real Upwork jobs.

Why does it matter?

There’s already a huge debate if Devin will replace software engineers. However, most production-grade software is too complex, unique, or domain-specific to be fully automated at this point. Perhaps, Devin could start handling more initial-level tasks in development. More so, it can assist developers in quickly prototyping, bootstrapping, and autonomously launching MVP for smaller apps and websites, for now

Source

Deepgram’s Aura empowers AI agents with authentic voices

Deepgram, a top voice recognition startup, just released Aura, its new real-time text-to-speech model. It’s the first text-to-speech model built for responsive, conversational AI agents and applications. Companies can use these agents for customer service in call centers and other customer-facing roles.

Deepgram’s Aura empowers AI agents with authentic voices
Deepgram’s Aura empowers AI agents with authentic voices

Aura includes a dozen natural, human-like voices with lower latency than any comparable voice AI alternative and is already being used in production by several customers. Aura works hand in hand with Deepgram’s Nova-2 speech-to-text API. Nova-2 is known for its top-notch accuracy and speed in transcribing audio streams.

Why does it matter?

Deepgram’s Aura is a one-stop shop for speech recognition and voice generation APIs that enable the fastest response times and most natural-sounding conversational flow. Its human-like voice models render extremely fast (typically in well under half a second) and at an affordable price ($0.015 per 1,000 characters). Lastly, Deepgram’s transcription is more accurate and faster than other solutions as well.

Source

Meta introduces two 24K GPU clusters to train Llama 3

Meta has invested significantly in its AI infrastructure by introducing two 24k GPU clusters. These clusters, built on top of Grand Teton, OpenRack, and PyTorch, are designed to support various AI workloads, including the training of Llama 3.

Meta introduces two 24K GPU clusters to train Llama 3
Meta introduces two 24K GPU clusters to train Llama 3

Meta aims to expand its infrastructure build-out by the end of 2024. It plans to include 350,000 NVIDIA H100 GPUs, providing compute power equivalent to nearly 600,000 H100s. The clusters are built with a focus on researcher and developer experience.

This adds up to Meta’s long-term vision to build open and responsibly developed artificial general intelligence (AGI). These clusters enable the development of advanced AI models and power applications such as computer vision, NLP, speech recognition, and image generation.

Why does it matter?

Meta is committed to open compute and open source, driving innovation in the AI software and hardware industry. Introducing two new GPUs to train Llama 3 is also a push forward to their commitment. As a founding member of Open Hardware Innovation (OHI) and the Open Innovation AI Research Community, Meta wants to make AI transparent and trustworthy.

Source

What Else Is Happening in AI on March 13th, 2024❗

🎮 Google Play to display AI-powered FAQs and recent YouTube videos for games

At the Google for Games Developer Summit held in San Francisco, Google announced several new features for ‘Google Play listing for games’. These include AI-powered FAQs, displaying the latest YouTube videos, new immersive ad formats, and support for native PC game publishing. These new features will allow developers to display promotions and the latest YouTube videos directly in their listing and show them to users in the Games tab of the Play Store. (Link)

🛡️ DoorDash’s new AI-powered tool automatically curbs verbal abuses

DoorDash has introduced a new AI-powered tool named ‘SafeChat+’ to review in-app conversations and determine if a customer or Dasher is being harassed. There will be an option to report the incident and either contact DoorDash’s support team if you’re a customer or quickly cancel the order if you’re a delivery person. With this feature, DoorDash aims to reduce verbally abusive and inappropriate interactions between consumers and delivery people. (Link)

🔍 Perplexity has decided to bring Yelp data to its chatbot

Perplexity has decided to bring Yelp data to its chatbot. The company CEO, Aravind Srinivas, told the media that many people use chatbots like search engines. He added that it makes sense to offer information on things they look for, like restaurants, directly from the source. That’s why they have decided to integrate Yelp’s maps, reviews, and other details in responses when people ask for restaurant or cafe recommendations.  (Link)

👗 Pinterest’s ‘body types ranges’ tool delivers more inclusive search results

Pinterest has introduced a new tool named body type ranges, which gives users a choice to self-select body types from a visual cue between four body type ranges to deliver personalized and more refined search results for women’s fashion and wedding inspiration. This tool aims to create a more inclusive place online to search, save, and shop. The company also plans to launch a similar feature for men’s fashion later this year. (Link)

🚀 OpenAI’s GPT-4.5 Turbo is all set to be launched in June 2024

According to the leak search engine results from Bing and DuckDuck Go, which indexed the OpenAI GPT-4.5 Turbo product page before an official announcement, OpenAI is all set to launch the new version of its LLM by June 2024. There is a discussion among the AI community that this could be OpenAI’s fastest, most accurate, and most scalable model to date. The details of GPT-4.5 Turbo were leaked by OpenAI’s web team, which now leads to a 404 page. (Link))

A Daily Chronicle of AI Innovations in March 2024 – Day 12: AI Daily News – March 12th, 2024

🚀Cohere’s introduces production-scale AI for enterprises
🤖 RFM-1 redefines robotics with human-like reasoning
🎧 Spotify introduces audiobook recommendations

🙃 Midjourney bans all its competitor’s employees

🚫 Google restricts election-related queries for its Gemini chatbot

📲 Apple to let developers distribute apps directly from their websites

💰 AI startups reach record funding of nearly $50 billion in 2023

Cohere’s introduces production-scale AI for enterprises

Cohere, an AI company, has introduced Command-R, a new large language model (LLM) designed to address real-world challenges, such as inefficient workflows, data analysis limitations, slow response times, etc.

Cohere’s introduces production-scale AI for enterprises
Cohere’s introduces production-scale AI for enterprises

Command-R focuses on two key areas: Retrieval Augmented Generation (RAG) and Tool Use. RAG allows the model to access and process information from private databases, improving the accuracy of its responses. Tool Use allows Command-R to interact with external software tools and APIs, automating complex tasks.

Command-R offers several features beneficial for businesses, including:

  • Multilingual capabilities: Supports 10 major languages
  • Cost-effectiveness: Offers a longer context window and reduced pricing compared to previous models
  • Wider accessibility: Available through Cohere’s API, major cloud providers, and free weights for research on HuggingFace

Overall, it empowers businesses to leverage AI for improved decision-making, increased productivity, and enhanced customer experiences.

Why does this matter?

Command-R showcases the future of business operations, featuring automated workflows, and enabling humans to focus on strategic work. Thanks to its low hallucination rate, we would see a wider adoption of AI technologies, and the development of sophisticated, context-aware AI applications tailored to specific business needs.

As AI continues to evolve and mature, models like Command-R will shape the future of work and the global economy.

Source

RFM-1 redefines robotics with human-like reasoning

Covariant has introduced RFM-1, a Robotics Foundation Model that gives robots ChatGPT-like understanding and reasoning capabilities.

TLDR;

  • RFM-1 is an 8 billion parameter transformer trained on text, images, videos, robot actions, and sensor readings from Covariant’s fleet of high-performing robotic systems deployed in real-world environments.
  • Similar to how we understand how objects move, RFM-1 can predict future outcomes/consequences based on initial images and robot actions.
  • RFM-1 leverages NLP to enable intuitive interfaces for programming robot behavior. Operators can instruct robots using plain English, lowering barriers to customizing AI behavior for specific needs.
  • RFM-1 can also communicate issues and suggest solutions to operators.

Why does this matter?

This advancement has the potential to revolutionize industries such as manufacturing, logistics, and healthcare, where robots can work alongside humans to improve efficiency, safety, and productivity.

Source

Spotify now recommends audiobooks (with AI)

Spotify has introduced a novel recommendation system called 2T-HGNN to provide personalized audiobook recommendations to its users. The system addresses the challenges of introducing a new content type (audiobooks) into an existing platform, such as data sparsity and the need for scalability.

Spotify now recommends audiobooks (with AI)
Spotify now recommends audiobooks (with AI)

2T-HGNN leverages a technique called “Heterogeneous Graph Neural Networks” (HGNNs) to uncover connections between different content types. Additionally, a “Two Tower” (2T) model helps ensure that recommendations are made quickly and efficiently for millions of users.

Interestingly, the system also uses podcast consumption data and weak interaction signals to uncover user preferences and predict future audiobook engagement.

Why does this matter?

This research will not only improve the user experience but also encourage users to explore and engage with audiobooks, potentially driving growth in this new content vertical. Moreover, it may inspire similar strategies in domains where tailored recommendations are essential, such as e-commerce, news, and entertainment.

Source

What Else Is Happening in AI on March 12th, 2024❗

💡 Elon Musk makes xAI’s Grok chatbot open-source

Elon Musk announced that his AI startup xAI will open-source its ChatGPT rival “Grok” this week, following a lawsuit against OpenAI for shifting to a for-profit model. Musk aims to provide free access to Grok’s code, aligning with open-source AI models like Meta and Mistral (Link)

 🖼️ Midjourney launches character consistent feature

Midjourney’s new “Consistent Character” feature lets artists create consistent characters across images. Users provide a reference image URL with their prompt, and the AI attempts to match the character’s features in new scenes. This holds promise for creators of comics, storyboards, and other visual narratives. (Link)

🤖 Apple tests AI for App Store ad optimization
Taking a page from Google and Meta, Apple is testing AI-powered ad placement within its App Store. This new system would automatically choose the most suitable locations (e.g., App Store Today page) to display ads based on advertiser goals and budget. This development could help Apple’s ad business reach $6 billion by 2025.(Link)

🏥China tests AI chatbot to assist neurosurgeons

China steps into the future of brain surgery with an AI co-pilot, dubbed “CARES Copilot”. This AI, based on Meta’s Llama 2.0, assists surgeons by analyzing medical data (e.g., scans) and offering informed suggestions during surgery. This government-backed project reflects China’s growing focus on developing domestic AI solutions for various sectors, including healthcare. (Link)

🧓South Korea deploys AI dolls to tackle elderly loneliness

Hyodol, a Korean-based company, has introduced an AI-powered companion doll to tackle loneliness among elderly. Priced at $1800, the robot doll boasts advanced features like conversation abilities, medication reminders, and safety alerts. With 7,000 dolls already deployed, Hyodol aims to expand to European and North American markets. (Link)

🙃 Midjourney bans all its competitor’s employees

  • Midjourney banned all Stability AI employees from using its service, citing a systems outage caused by data scraping efforts linked to Stability AI employees.
  • The company announced the ban and a new policy against “aggressive automation” after identifying botnet-like activity from Stability AI during a server outage.
  • Stability AI CEO Emad Mostaque is looking into the incident, and Midjourney’s founder David Holz has provided information for the internal investigation.
  • Source

🚫 Google restricts election-related queries for its Gemini chatbot

  • Google has begun restricting Gemini queries related to elections globally in countries where elections are taking place, to prevent the dissemination of false or misleading information.
  • The restrictions were implemented amid concerns over generative AI’s potential impact on elections and followed an advisory from India requiring tech firms to obtain government permission before introducing new AI models.
  • Despite the restrictions, the effectiveness of the restrictions is under question as some users found ways to bypass them, and it’s uncertain if Google will lift these restrictions post-elections.
  • Source

💰 AI startups reach record funding of nearly $50 billion in 2023

  • AI startups reached a record funding of nearly $50 billion in 2023, with significant contributions from companies like OpenAI and Anthropic.
  • Investment trends showed over 70 funding rounds exceeding $100 million each, partly due to major companies’ investments, including Microsoft’s $10 billion in OpenAI.
  • While large tech companies are venturing to dominate the AI market, specialized AI startups like Midjourney manage to maintain niches by offering superior products.
  • Source

A Daily Chronicle of AI Innovations in March 2024 – Day 11: AI Daily News – March 11th, 2024

🖼️ Huawei’s PixArt-Σ paints prompts to perfection
🧠 Meta cracks the code to improve LLM reasoning
📈 Yi Models exceed benchmarks with refined data

Huawei’s PixArt-Σ paints prompts to perfection

Researchers from Huawei’s Noah’s Ark Lab introduced PixArt-Σ, a text-to-image model that can create 4K resolution images with impressive accuracy in following prompts. Despite having significantly fewer parameters than models like SDXL, PixArt-Σ outperforms them in image quality and prompt matching.

  

The model uses a “weak-to-strong” training strategy and efficient token compression to reduce computational requirements. It relies on carefully curated training data with high-resolution images and accurate descriptions, enabling it to generate detailed 4K images closely matching the text prompts. The researchers claim that PixArt-Σ can even keep up with commercial alternatives such as Adobe Firefly 2, Google Imagen 2, OpenAI DALL-E 3, and Midjourney v6.

Why does this matter?

PixArt-Σ’s ability to generate high-resolution, photorealistic images accurately could impact industries like advertising, media, and entertainment. As its efficient approach requires fewer computational resources than existing models, businesses may find it easier and more cost-effective to create custom visuals for their products or services.

Source

Meta cracks the code to improve LLM reasoning

Meta researchers investigated using reinforcement learning (RL) to improve the reasoning abilities of large language models (LLMs). They compared algorithms like Proximal Policy Optimization (PPO) and Expert Iteration (EI) and found that the simple EI method was particularly effective, enabling models to outperform fine-tuned models by nearly 10% after several training iterations.

However, the study also revealed that the tested RL methods have limitations in further improving LLMs’ logical capabilities. The researchers suggest that stronger exploration techniques, such as Tree of Thoughts, XOT, or combining LLMs with evolutionary algorithms, are important for achieving greater progress in reasoning performance.

Why does this matter?

Meta’s research highlights the potential of RL in improving LLMs’ logical abilities. This could lead to more accurate and efficient AI for domains like scientific research, financial analysis, and strategic decision-making. By focusing on techniques that encourage LLMs to discover novel solutions and approaches, researchers can make more advanced AI systems.

Source

Yi models exceed benchmarks with refined data

01.AI has introduced the Yi model family, a series of language and multimodal models that showcase impressive multidimensional abilities. The Yi models, based on 6B and 34B pretrained language models, have been extended to include chat models, 200K long context models, depth-upscaled models, and vision-language models.

The performance of the Yi models can be attributed to the high-quality data resulting from 01.AI‘s data-engineering efforts. By constructing a massive 3.1 trillion token dataset of English and Chinese corpora and meticulously polishing a small-scale instruction dataset, 01.AI has created a solid foundation for their models. The company believes that scaling up model parameters using thoroughly optimized data will lead to even more powerful models.

Why does this matter?

The Yi models’ success in language, vision, and multimodal tasks suggests that they could be adapted to a wide range of applications, from customer service chatbots to content moderation and beyond. These models also serve as a prime example of how investing in data optimization can lead to groundbreaking advancements in the field.

Source

OpenAI’s Evolution into Skynet: AI and Robotics Future, Figure Humanoid Robots

 

  • OpenAI’s partnership with Figure signifies a transformative step in the evolution of AI and robotics.
  • Utilizing Microsoft Azure, OpenAI’s investment supports the deployment of autonomous humanoid robots for commercial use.
  • Figure’s collaboration with BMW Manufacturing integrates humanoid robots to enhance automotive production.
  • This technological progression echoes the fictional superintelligence Skynet yet emphasizes real-world innovation and safety.
  • The industry valuation of Figure at $2.6 billion underlines the significant impact and potential of advanced AI in commercial sectors.

What Else Is Happening in AI on March 11, 2024❗

🏠 Redfin’s AI can tell you about your dream neighborhood

“Ask Redfin” can now answer questions about homes, neighborhoods, and more. Using LLMss, the chatbot can provide insights on air conditioning, home prices, safety, and even connect users to agents. It is currently available in 12 U.S. cities, including Atlanta, Boston, Chicago, and Washington, D.C. (Link)

🔊 Pika Labs Adds Sound to Silent AI Videos 

Pika Labs users can now add sound effects to their generated videos. Users can either specify the exact sounds they want or let Pika’s AI automatically select and integrate them based on the video’s content. This update aims to provide a more immersive and engaging video creation experience, setting a new standard in the industry. (Link)

🩺 Salesforce’s new AI tool for doctors automates paperwork

Salesforce is launching new AI tools to help healthcare workers automate tedious administrative tasks. Einstein Copilot: Health Actions will allow doctors to book appointments, summarize patient info, and send referrals using conversational AI, while Assessment Generation will digitize health assessments without manual typing or coding. (Link)

🖥️ HP’s new AI-powered PCs redefine work 

HP just dropped a massive lineup of AI-powered PCs, including the HP Elite series, Z by HP mobile workstations, and Poly Studio conferencing solutions. These devices use AI to improve productivity, creativity, and collaboration for the hybrid workforce, while also offering advanced security features like protection against quantum computer hacks. (Link)

🎨 DALL-E 3’s new look is artsy and user-friendly

OpenAI is testing a new user interface for DALL-E 3. It allows users to choose between predefined styles and aspect ratios directly in the GPT, offering a more intuitive and educational experience. OpenAI has also implemented the C2PA standard for metadata verification and is working on an image classifier to reliably recognize DALL-E images. (Link)

A Daily Chronicle of AI Innovations in March 2024 – Week 1 Summary

  1. Anthropic introduced the next generation of Claude: Claude 3 model family. It includes OpusSonnet and Haiku models. Opus is the most intelligent model, that outperforms GPT-4 and Gemini 1.0 Ultra on most of the common evaluation benchmarks. Haiku is the fastest, most compact model for near-instant responsiveness. The Claude 3 models have vision capabilities, offer a 200K context window capable of accepting inputs exceeding 1 million tokens, improved accuracy and fewer refusals [Details | Model Card].
  2. Stability AI partnered with Tripo AI and released TripoSR, a fast 3D object reconstruction model that can generate high-quality 3D models from a single image in under a second. The model weights and source code are available under the MIT license, allowing commercialized use. [Details | GitHub | Hugging Face].
  3. Answer.AI released a fully open source system that, for the first time, can efficiently train a 70b large language model on a regular desktop computer with two or more standard gaming GPUs. It combines QLoRA with Meta’s FSDP, which shards large models across multiple GPUs [Details].
  4. Inflection launched Inflection-2.5, an upgrade to their model powering Pi, Inflection’s empathetic and supportive companion chatbot. Inflection-2.5 approaches GPT-4’s performance, but used only 40% of the amount of compute for training. Pi is also now available on Apple Messages [Details].
  5. Twelve Labs introduced Marengo-2.6, a new state-of-the-art (SOTA) multimodal foundation model capable of performing any-to-any search tasks, including Text-To-Video, Text-To-Image, Text-To-Audio, Audio-To-Video, Image-To-Video, and more [Details].
  6. Cloudflare announced the development of Firewall for AI, a protection layer that can be deployed in front of Large Language Models (LLMs), hosted on the Cloudflare Workers AI platform or models hosted on any other third party infrastructure, to identify abuses before they reach the models [Details]
  7. Scale AI, in partnership with the Center for AI Safety, released WMDP (Weapons of Mass Destruction Proxy): an open-source evaluation benchmark of 4,157 multiple-choice questions that serve as a proxy measurement of LLM’s risky knowledge in biosecurity, cybersecurity, and chemical security [Details].
  8. Midjourney launched v6 turbo mode to generate images at 3.5x the speed (for 2x the cost). Just type /turbo [Link].
  9. Moondream.ai released moondream 2 – a small 1.8B parameters, open-source, vision language model designed to run efficiently on edge devices. It was initialized using Phi-1.5 and SigLIP, and trained primarily on synthetic data generated by Mixtral. Code and weights are released under the Apache 2.0 license, which permits commercial use [Details].
  10. Vercel released Vercel AI SDK 3.0. Developers can now associate LLM responses to streaming React Server Components [Details].
  11. Nous Research released a new model designed exclusively to create instructions from raw-text corpuses, Genstruct 7B. This enables the creation of new, partially synthetic instruction finetuning datasets from any raw-text corpus [Details].
  12. 01.AI open-sources Yi-9B, one of the top performers among a range of similar-sized open-source models excelling in code, math, common-sense reasoning, and reading comprehension [Details].
  13. Accenture to acquire Udacity to build a learning platform focused on AI [Details].
  14. China Offers ‘Computing Vouchers’ upto $280,000 to Small AI Startups to train and run large language models [Details].
  15. Snowflake and Mistral have partnered to make Mistral AI’s newest and most powerful model, Mistral Large, available in the Snowflake Data Cloud [Details]
  16. OpenAI rolled out ‘Read Aloud’ feature for ChatGPT, enabling ChatGPT to read its answers out loud. Read Aloud can speak 37 languages but will auto-detect the language of the text it’s reading [Details].

A Daily Chronicle of AI Innovations in March 2024 – Day 8: AI Daily News – March 08th, 2024

🗣️Inflection 2.5: A new era of personal AI is here!
🔍Google announces LLMs on device with MediaPipe
🤖GaLore: A new method for memory-efficient LLM training

📱Adobe makes creating social content on mobile easier

🛡️OpenAI now allows users to add MFA to user accounts

🏅US Army is building generative AI chatbots in war games

🧑‍🎨 Claude 3 builds the painting app in 2 minutes and 48 seconds

🧪Cognizant launches AI lab in San Francisco to drive innovation

Inflection 2.5: A new era of personal AI is here!

Inflection.ai, the company behind the personal AI app Pi, has recently introduced Inflection-2.5, an upgraded large language model (LLM) that competes with top LLMs like GPT-4 and Gemini. The in-house upgrade offers enhanced capabilities and improved performance, combining raw intelligence with the company’s signature personality and empathetic fine-tuning.

Inflection 2.5: A new era of personal AI is here!
Inflection 2.5: A new era of personal AI is here!

This upgrade has made significant progress in coding and mathematics, keeping Pi at the forefront of technological innovation. With Inflection-2.5, Pi has world-class real-time web search capabilities, providing users with high-quality breaking news and up-to-date information. This empowers Pi users with a more intelligent and empathetic AI experience.

Why does it matter?

Inflection-2.5 challenges leading language models like GPT-4 and Gemini with their raw capability, signature personality, and empathetic fine-tuning. This will provide a new alternative for startups and enterprises building personalized applications with generative AI capabilities.

Source

Google announces LLMs on device with MediaPipe

Google’s new experimental release called the MediaPipe LLM Inference API  allows LLMs to run fully on-device across platforms. This is a significant development considering LLMs’ memory and computing demands, which are over a hundred times larger than traditional on-device models.

Google announces LLMs on device with MediaPipe
Google announces LLMs on device with MediaPipe

The MediaPipe LLM Inference API is designed to streamline on-device LLM integration for web developers and supports Web, Android, and iOS platforms. It offers several key features and optimizations that enable on-device AI. These include new operations, quantization, caching, and weight sharing. Developers can now run LLMs on devices like laptops and phones using MediaPipe LLM Inference API.

Why does it matter?

Running LLMs on devices using MediaPipe and TensorFlow Lite allows for direct deployment, reducing dependence on cloud services. On-device LLM operation ensures faster and more efficient inference, which is crucial for real-time applications like chatbots or voice assistants. This innovation helps rapid prototyping with LLM models and offers streamlined platform integration.

Source

GaLore: A new method for memory-efficient LLM training

Researchers have developed a new technique called Gradient Low-Rank Projection (GaLore) to reduce memory usage while training large language models significantly. Tests have shown that GaLore achieves results similar to full-rank training while reducing optimizer state memory usage by up to 65.5% when pre-training large models like LLaMA.

GaLore: A new method for memory-efficient LLM training
GaLore: A new method for memory-efficient LLM training

It also allows pre-training a 7 billion parameter model from scratch on a single 24GB consumer GPU without needing extra techniques. This approach works well for fine-tuning and outperforms low-rank methods like LoRA on GLUE benchmarks while using less memory. GaLore is optimizer-independent and can be used with other techniques like 8-bit optimizers to save additional memory.

Why does it matter?

The gradient matrix’s low-rank nature will help AI developers during model training. GaLore minimizes the memory cost of storing gradient statistics for adaptive optimization algorithms. It enables training large models like LLaMA with reduced memory consumption, making it more accessible and efficient for researchers.

Source

🤖 OpenAI CTO complained to board about ‘manipulative’ CEO Sam Altman 

  • OpenAI CTO Mira Murati was reported by the New York Times to have played a significant role in CEO Sam Altman’s temporary removal, raising concerns about his leadership in a private memo and with the board.
  • Altman was accused of creating a toxic work environment, leading to fears among board members that key executives like Murati and co-founder Ilya Sutskever could leave, potentially causing a mass exit of talent.
  • Despite internal criticisms of Altman’s leadership and management of OpenAI’s startup fund, hundreds of employees threatened to leave if he was not reinstated, highlighting deep rifts within the company’s leadership.
  • Source

Saudi Arabia’s Male Humanoid Robot Accused of Sexual Harassment

A video of Saudi Arabia’s first male robot has gone viral after a few netizens accused the humanoid of touching a female reporter inappropriately.

“Saudi Arabia unveils its man-shaped AI robot, Mohammad, reacts to a reporter in its first appearance,” an X user wrote while sharing the video that people are claiming shows the robot’s inappropriate behaviour. You can view the original tweet here.

What Else Is Happening in AI on March 08th, 2024❗

📱Adobe makes creating social content on mobile easier

Adobe has launched an updated version of Adobe Express, a mobile app that now includes Firefly AI models. The app offers features such as a “Text to Image” generator, a “Generative Fill” feature, and a “Text Effects” feature, which can be utilized by small businesses and creative professionals to enhance their social media content. Creative Cloud members can also access and work on creative assets from Photoshop and Illustrator directly within Adobe Express. (Link)

🛡️OpenAI now allows users to add MFA to user accounts

To add extra security to OpenAI accounts, users can now enable Multi-Factor Authentication (MFA). To set up MFA, users can follow the instructions in the OpenAI Help Center article “Enabling Multi-Factor Authentication (MFA) with OpenAI.” MFA requires a verification code with their password when logging in, adding an extra layer of protection against unauthorized access. (Link)

🏅US Army is building generative AI chatbots in war games

The US Army is experimenting with AI chatbots for war games. OpenAI’s technology is used to train the chatbots to provide battle advice. The AI bots act as military commanders’ assistants, offering proposals and responding within seconds. Although the potential of AI is acknowledged, experts have raised concerns about the risks involved in high-stakes situations. (Link)

🧑‍🎨 Claude 3 builds the painting app in 2 minutes and 48 seconds

Claude 3, the latest AI model by Anthropic, created a multiplayer drawing app in just 2 minutes and 48 seconds. Multiple users could collaboratively draw in real-time with user authentication and database integration. The AI community praised the app, highlighting the transformative potential of AI in software development. Claude 3 could speed up development cycles and make software creation more accessible. (Link)

🧪Cognizant launches AI lab in San Francisco to drive innovation

Cognizant has opened an AI lab in San Francisco to accelerate AI adoption in businesses. The lab, staffed with top researchers and developers, will focus on innovation, research, and developing cutting-edge AI solutions. Cognizant’s investment in AI research positions them as a thought leader in the AI space, offering advanced solutions to meet the modernization needs of global enterprises. (Link)

A Daily Chronicle of AI Innovations in March 2024 – Day 7: AI Daily News – March 07th, 2024

🗣️Microsoft’s NaturalSpeech makes AI sound human
🔍Google’s search update targets AI-generated spam
🤖Google’s RT-Sketch teaches robots with doodles

🕵️ Ex-Google engineer charged with stealing AI secrets for Chinese firm

🚨 Microsoft engineer sounds alarm on company’s AI image generator in letter to FTC

🤔 Apple bans Epic’s developer account and calls the company ‘verifiably untrustworthy’

🍎 Apple reportedly developing foldable MacBook with 20.3-inch screen

🧠 Meta is building a giant AI model to power its ‘entire video ecosystem

Microsoft’s NaturalSpeech makes AI sound human

Microsoft and its partners have created NaturalSpeech 3, a new Text-to-Speech system that makes computer-generated voices sound more human. Powered by FACodec architecture and factorized diffusion models, NaturalSpeech 3 breaks down speech into different parts, like content, tone, and sound quality to create a natural-sounding speech that fits specific prompts, even for voices it hasn’t heard before.

Microsoft's NaturalSpeech makes AI sound human
Microsoft’s NaturalSpeech makes AI sound human

NaturalSpeech 3 works better than other voice tech in terms of quality, similarity, tone, and clarity. It keeps getting better as it learns from more data. By letting users change how the speech sounds through prompts, NaturalSpeech 3 makes talking to computers feel more like talking to a person. This research is a big step towards a future where chatting with computers is as easy as chatting with friends.

Why does this matter?

This advancement transcends mere voice quality. This could change the way we interact with devices like smartphones, smart speakers, and virtual assistants. Imagine having a more natural, engaging conversation with Siri, Alexa, or other AI helpers.

Better voice tech could also make services more accessible for people with visual impairments or reading difficulties. It might even open up new possibilities in entertainment, like more lifelike characters in video games or audiobooks that sound like they’re read by your favorite celebrities.

Source

Google’s search update targets AI-generated spam

Google has announced significant changes to its search ranking algorithms in order to reduce low-quality and AI-generated spam content in search results. The March update targets three main spam practices: mass distribution of unhelpful content, abusing site reputation to host low-quality content, and repurposing expired domains with poor content.

While Google is not devaluing all AI-generated content, it aims to judge content primarily on its usefulness to users. Most of the algorithm changes are effective immediately, though sites abusing their reputation have a 60-day grace period to change their practices. As Google itself develops AI tools, SGE and Gemini, the debate around AI content and search result quality is just beginning.

Why does this matter?

Websites that churn out lots of AI-made content to rank higher on Google may see their rankings drop. This might push them to focus more on content creation strategies, with a greater emphasis on quality over quantity.

For people using Google, the changes should mean finding more useful results and less junk.

As AI continues to advance, search engines like Google will need to adapt their algorithms to surface the most useful content, whether it’s written by humans or AI.

Source

Google’s RT-Sketch teaches robots with doodles

Google has introduced RT-Sketch, a new approach to teaching robots tasks using simple sketches. Users can quickly draw a picture of what they want the robot to do, like rearranging objects on a table. RT-Sketch focuses on the essential parts of the sketch, ignoring distracting details.

Google's RT-Sketch teaches robots with doodles
Google’s RT-Sketch teaches robots with doodles

Source

RT-Sketch is trained on a dataset of paired trajectories and synthetic goal sketches, and tested on six object rearrangement tasks. The results show that RT-Sketch performs comparably to image or language-conditioned agents in simple settings with written instructions on straightforward tasks. However, it did better when instructions were confusing or there were distracting objects present.

RT-Sketch can also interpret and act upon sketches with varying levels of detail, from basic outlines to colorful drawings.

Why does this matter?

With RT-Sketch, people can tell robots what to do without needing perfect images or detailed written instructions. This could make robots more accessible and useful in homes, workplaces, and for people who have trouble communicating in other ways.

As robots become a bigger part of our lives, easy ways to talk to them, like sketching, could help us get the most out of them. RT-Sketch is a step toward making robots that better understand what we need.

Source

What Else Is Happening in AI on March 07th, 2024❗

🤖Google’s Gemini lets users edit within the chatbox

Google has updated its Gemini chatbot, allowing users to directly edit and fine-tune responses within the chatbox. This feature, launched on March 4th for English users in the Gemini web app, enables more precise outputs by letting people select text portions and provide instructions for improvement. (Link)

📈Adobe’s AI boosts IBM’s marketing efficiency

IBM reports a 10-fold increase in designer productivity and a significant reduction in marketing campaign time after testing Adobe’s generative AI tools. The AI-powered tools have streamlined idea generation and variant creation, allowing IBM to achieve more in less time. (Link)

💡 Zapier’s new tool lets you make AI bots without coding

Zapier has released Zapier Central, a new AI tool that allows users to create custom AI bots by simply describing what they want, without any coding. The bots can work with Zapier’s 6,000+ connected apps, making it easy for businesses to automate tasks. (Link)

🤝Accenture teams up with Cohere to bring AI to enterprises

Accenture has partnered with AI startup, Cohere to provide generative AI solutions to businesses. Leveraging Cohere’s language models and search technologies, the collaboration aims to boost productivity and efficiency while ensuring data privacy and security. (Link)

🎥 Meta builds mega AI model for video recommendations
Meta is developing a single AI model to power its entire video ecosystem across platforms by 2026. The company has invested billions in Nvidia GPUs to build this model, which has already shown promising results in improving Reels watch time on the core Facebook app. (Link)

OpenAI is researching photonic processors to run their AI on

OpenAI hired this person:  He has been doing a lot of research on waveguides for photonic processing for both Training AI and for inference and he did a PHD about photonic waveguides:

I think that he is going to help OpenAI to build photonic waveguides that they can run their neural networks / AI Models on and this is really  cool if OpenAI actually think that they can build processors with faster Inference and training with photonics.

🕵️ Ex-Google engineer charged with stealing AI secrets for Chinese firm

  • Linwei Ding, a Google engineer, has been indicted for allegedly stealing over 500 files related to Google’s AI technology, including designs for chips and data center technologies, to benefit companies in China.
  • The stolen data includes designs for Google’s TPU chips and GPUs, crucial for AI workloads, amid U.S. efforts to restrict China’s access to AI-specific chips.
  • Ding allegedly transferred stolen files to a personal cloud account using a method designed to evade Google’s detection systems, was offered a CTO position by a Chinese AI company and founded a machine learning startup in China while still employed at Google.
  • Source

🚨 Microsoft engineer sounds alarm on company’s AI image generator in letter to FTC

  • Microsoft AI engineer Shane Jones warns that the company’s AI image generator, Copilot Designer, generates sexual and violent content and ignores copyright laws.
  • Jones shared his findings with Microsoft and contacted U.S. senators and the FTC, demanding better safeguards and an independent review of Microsoft’s AI incident reporting process.
  • In addition to the problems with Copilot Designer, other Microsoft products based on OpenAI technologies, such as Copilot Chat, tend to have poorer performance and more insecure implementations than the original OpenAI products, such as ChatGPT and DALL-E 3.
  • Source

🧠 Meta is building a giant AI model to power its ‘entire video ecosystem’ 

  • Meta is developing an AI model designed to power its entire video ecosystem, including the TikTok-like Reels service and traditional video content, as part of its technology roadmap through 2026.
  • The company has invested billions of dollars in Nvidia GPUs to support this AI initiative, aiming to improve recommendation systems and overall product performance across all platforms.
  • This AI model has already demonstrated an 8% to 10% increase in Reels watch time on the Facebook app, with Meta now working to expand its application to include the Feed recommendation product and possibly integrate sophisticated chatting tools.
  • Innovating for the Future

    As Meta continues to innovate and refine their AI model architecture, we can expect even more exciting developments in the future. The company’s dedication to enhancing the video recommendation experience and leveraging the full potential of AI is paving the way for a new era in online video consumption.

    Stay tuned for more updates as Meta strives to revolutionize the digital video landscape with its cutting-edge AI technology.

    r/aidailynewsupdates - Meta's AI Model to Revolutionize Video Ecosystem
  • Source

Will AI destroy the adtech industry?

Some points to consider on both sides:

Yes:

– AI will enable humans to get content they want, nothing more

– New AI OSes will act ‘for’ the human, cleaning content of ads

– OpenAI and new startups don’t need ad revenue, they’ll take monthly subscriptions to deliver information with no ads

No:

– New AI OSes will integrate ads even more closely into the computing experience, acting ‘against’ the human

– Content will be more tightly integrated with ads, and AI won’t be able to unpiece this

– Meta and Alphabet have $100bns of skin in the game, they will make sure this doesn’t happen, including by using their lawyers to prevent lifting content out of the ad context

A Daily Chronicle of AI Innovations in March 2024 – Day 6: AI Daily News – March 06th, 2024

🏆 Microsoft’s Orca AI beats 10x bigger models in math
🎨 GPT-4V wins at turning designs into code
🎥 DeepMind alums’ Haiper joins the AI video race

🤔 OpenAI fires back, says Elon Musk demanded ‘absolute control’ of the company

📱 iOS 17.4 is here: what you need to know

🚫 TikTok faces US ban if ByteDance fails to sell app

🔍 Google now wants to limit the AI-powered search spam it helped create

OpenAI vs Musk (openai responds to elon musk).

 What does Elon mean by: “Unfortunately, humanity’s future is in the hands of <redacted>”? Is it google?

What does elon mean "Unfortunately, humanity's future is in the hands of <redacted>"? Is it google?
What does elon mean “Unfortunately, humanity’s future is in the hands of “? Is it google?

What does elon mean "Unfortunately, humanity's future is in the hands of <redacted>"? Is it google?
What does elon mean “Unfortunately, humanity’s future is in the hands of “? Is it google?

  • OpenAI has countered Elon Musk’s lawsuit by revealing Musk’s desire for “absolute control” over the company, including merging it with Tesla, holding majority equity, and becoming CEO.
  • In a blog post, OpenAI aims to dismiss Musk’s claims and argues against his view that the company has deviated from its original nonprofit mission and has become too closely aligned with Microsoft.
  • OpenAI defends its stance on not open-sourcing its work, citing a 2016 email exchange with Musk that supports a less open approach as the development of artificial general intelligence advances.

For the first time in history, an AI has a higher IQ than the average human.

For the first time in history, an AI has a higher IQ than the average human.
For the first time in history, an AI has a higher IQ than the average human.

Claude 3 vs. GPT-4

Right now, the question on everyone’s mind is whether Claude 3 is better than GPT-4. It’s a fair question; GPT-4 has dominated the LLM benchmarks for over a year, despite plenty of competitors trying to catch up.

Certainly, GPT-4 now has some real competition in the form of Claude 3 and Gemini 1.5. Even if we put the benchmarks aside for a moment, capabilities like video comprehension and million-token context windows are pushing the state of the art forward, and OpenAI could finally cede its dominant position.

But I think that “best,” when it comes to LLMs, is a little bit of a red herring. Despite the marketing and social media hype, these models have more similarities than differences. Ultimately, “best” depends on your use cases and preferences.

Claude 3 may be better at reasoning and language comprehension than GPT-4, but that won’t matter much if you’re mainly generating code. Likewise, Gemini 1.5 may have better multi-modal capabilities, but if you’re concerned with working in different languages, then Claude might be your best bet. In my (very limited) testing, I’ve found that Opus is a much better writer than GPT-4 – the default writing style is far more “normal” than what I can now recognize as ChatGPT-generated content. But I’ve yet to try brainstorming and code generation tasks.

So, for now, my recommendation is to keep experimenting and find a model that works for you. Not only because each person’s use cases differ but also because the models are regularly improving! In the coming months, Anthropic plans to add function calls, interactive coding, and more agentic capabilities to Claude 3.

To try Claude 3 for yourself, you can start talking with Claude 3 Sonnet today (though you’ll need to be in one of Anthropic’s supported countries). Opus is available to paid subscribers of Claude Pro. If you’re a developer, Opus and Sonnet are available via the API, and Sonnet is additionally available through Amazon Bedrock and Google Cloud’s Vertex AI Model Garden. The models are also available via a growing number of third-party apps and services: check your favorite AI tool to see if it supports Claude 3!

Guy builds an AI-steered homing/killer drone in just a few hours

Guy builds an AI-steered homing/killer drone in just a few hours
Guy builds an AI-steered homing/killer drone in just a few hours

Read Aloud For Me AI Dashboard on the App Store (apple.com)

Always Say Hello to Your GPTs… (Better Performing Custom GPTs)

I’ve been testing out lots of custom GPTs that others have made. Specifically games and entertaining GPTs and I noticed some issues and a solution.

The problem: First off, many custom GPT games seem to forget to generate images as per their instructions. I also noticed that, often, the game or persona (or whatever the GPT aims to be) becomes more of a paraphrased or simplified version of what it should be and responses become more like base ChatGPT.

The solution: I’ve noticed that custom GPTs will perform much better if the user starts the initial conversation with a simple ”Hello, can you explain your functionality and options to me?”. This seems to remind the custom GPT of it’s tone ensures it follow’s its instructions.

Microsoft’s Orca AI beats 10x bigger models in math

Microsoft’s Orca team has developed Orca-Math, an AI model that excels at solving math word problems despite its compact size of just 7 billion parameters. It outperforms models ten times larger on the GSM8K benchmark, achieving 86.81% accuracy without relying on external tools or tricks. The model’s success is attributed to training on a high-quality synthetic dataset of 200,000 math problems created using multi-agent flows and an iterative learning process involving AI teacher and student agents.

Microsoft's Orca AI beats 10x bigger models in math
Microsoft’s Orca AI beats 10x bigger models in math

The Orca team has made the dataset publicly available under the MIT license, encouraging researchers and developers to innovate with the data. The small dataset size highlights the potential of using multi-agent flows to generate data and feedback efficiently.

Why does this matter?

Orca-Math’s breakthrough performance shows the potential for smaller, specialized AI models in niche domains. This development could lead to more efficient and cost-effective AI solutions for businesses, as smaller models require less computational power and training data, giving companies a competitive edge.

Source

GPT-4V wins at turning designs into code

With unprecedented capabilities in multimodal understanding and code generation, GenAI can enable a new paradigm of front-end development where LLMs directly convert visual designs into code implementation. New research formalizes this as “Design2Code” task and conduct comprehensive benchmarking. It also:

  • Introduces Design2Code benchmark consisting of diverse real-world webpages as test examples
  • Develops comprehensive automatic metrics that complement human evaluations
  • Proposes new multimodal prompting methods that improve over direct prompting baselines.
  • Finetunes open-source Design2Code-18B model that matches the performance of Gemini Pro Vision on both human and automatic evaluation

Moreover, it finds 49% of the GPT-4V-generations webpages were good enough to replace the original references, while 64% were even better designed than the original references.

Why does this matter?

This research could simplify web development for anyone to build websites from visual designs using AI, much like word processors made writing accessible. For enterprises, automating this front-end coding process could improve collaboration between teams and speed up time-to-market across industries if implemented responsibly alongside human developers.

Source

What Else Is Happening in AI on March 06th, 2024❗

📸 Kayak’s AI finds cheaper flights from screenshots

Kayak introduced two new AI features: PriceCheck, which lets users upload flight screenshots to find cheaper alternatives and Ask Kayak, a ChatGPT-powered travel advice chatbot. These additions position Kayak alongside other travel sites, using generative AI to improve trip planning and flight price comparisons in a competitive market. (Link)

🎓 Accenture invests $1B in LearnVantage for AI upskilling

Accenture is launching LearnVantage, investing $1 billion over three years to provide clients with customized technology learning and training services. Accenture is also acquiring Udacity to scale its learning capabilities and meet the growing demand for technology skills, including generative AI, so organizations can achieve business value using AI. (Link)

🤝 Snowflake brings Mistral’s LLMs to its data cloud

Snowflake has partnered with Mistral AI to bring Mistral’s open LLMs into its Data Cloud. This move allows Snowflake customers to build LLM apps directly within the platform. It also marks a significant milestone for Mistral AI, which has recently secured partnerships with Microsoft, IBM, and Amazon. The deal positions Snowflake to compete more effectively in the AI space and increases Mistral AI visibility. (Link)

🛡️ Dell & CrowdStrike unite to fight AI threats

Dell and CrowdStrike are partnering to help businesses fight cyberattacks using AI. By integrating CrowdStrike’s Falcon XDR platform into Dell’s MDR service, they aim to protect customers against threats like generative AI attacks, social engineering, and endpoint breaches. (Link)

📱 AI app diagnoses ear infections with a snap

Physician-scientists at UPMC and the University of Pittsburgh have developed a smartphone app that uses AI to accurately diagnose ear infections (acute otitis media) in young children. The app analyzes short videos of the eardrum captured by an otoscope connected to a smartphone camera. It could help decrease unnecessary antibiotic use by providing a more accurate diagnosis than many clinicians. (Link)

DeepMind alums’ Haiper joins the AI video race

DeepMind alums Yishu Miao and Ziyu Wang have launched Haiper, a video generation tool powered by their own AI model. The startup offers a free website where users can generate short videos using text prompts, although there are limitations on video length and quality.

DeepMind alums' Haiper joins the AI video race
DeepMind alums’ Haiper joins the AI video race

The company has raised $19.2 million in funding and focuses on improving its AI model to deliver high-quality, realistic videos. They aim to build a core video generation model that can be offered to developers and address challenges like the “uncanny valley” problem in AI-generated human figures.

Why does this matter?

Haiper signals the race to develop video AI models that can disrupt industries like marketing, entertainment, and education by allowing businesses to generate high-quality video content cost-effectively. However, the technology is at an early stage, so there is room for improvement, highlighting the need for responsible development.

Source

A Daily Chronicle of AI Innovations in March 2024 – Day 5: AI Daily News – March 05th, 2024

🏆Anthropic’s Claude 3 Beats OpenAI’s GPT-4
🖼️ TripsoSR: 3D object generation from a single image in <1s
🔒 Cloudflare’s Firewall for AI protects LLMs from abuses

🥴 Google co-founder says company ‘definitely messed up’

🚫 Facebook, Instagram, and Threads are all down

🤔 Microsoft compares New York Times to ’80s movie studios trying to ban VCRs

💼 Fired Twitter execs are suing Elon Musk for over $128 million

Claude 3 gets ~60% accuracy on GPQA

 Claude 3 gets ~60% accuracy on GPQA
Claude 3 gets ~60% accuracy on GPQA

Anthropic’s Claude 3 beats OpenAI’s GPT-4

Anthropic has launched Claude 3, a new family of models that has set new industry benchmarks across a wide range of cognitive tasks. The family comprises three state-of-the-art models in ascending order of cognitive ability: Claude 3 Haiku, Claude 3 Sonnet, and Claude 3 Opus. Each model provides an increasing level of performance, and you can choose the one according to your intelligence, speed, and cost requirements.

Anthropic’s Claude 3 beats OpenAI’s GPT-4
Anthropic’s Claude 3 beats OpenAI’s GPT-4

Opus and Sonnet are now available via claude.ai and the Claude API in 159 countries, and Haiku will join that list soon.

Claude 3 has set a new standard of intelligence among its peers on most of the common evaluation benchmarks for AI systems, including undergraduate-level expert knowledge (MMLU), graduate-level expert reasoning (GPQA), basic mathematics (GSM8K), and more.

Anthropic’s Claude 3 beats OpenAI’s GPT-4
Anthropic’s Claude 3 beats OpenAI’s GPT-4

In addition, Claude 3 displays solid visual processing capabilities and can process a wide range of visual formats, including photos, charts, graphs, and technical diagrams.  Lastly, compared to Claude 2.1, Claude 3 exhibits 2x accuracy and precision for responses and correct answers.

Why does it matter?

In 2024, Gemini and ChatGPT caught the spotlight, but now Claude 3 has emerged as the leader in AI benchmarks. While benchmarks matter, only the practical usefulness of Claude 3 will tell if it is truly superior. This might also prompt OpenAI to release a new ChatGPT upgrade. However, with AI models becoming more common and diverse, it’s unlikely that one single model will emerge as the ultimate winner.

Source

TripsoSR: 3D object generation from a single image in <1s

Stability AI has introduced a new AI model named TripsoSR in partnership with Trip AI. The model enables high-quality 3D object generation or rest from a single in less than a second. It runs under low inference budgets (even without a GPU) and is accessible to many users.

TripsoSR: 3D object generation from a single image in <1s
TripsoSR: 3D object generation from a single image in <1s

As far as performance, TripoSR can create detailed 3D models in a fraction of the time of other models. When tested on an Nvidia A100, it generates draft-quality 3D outputs (textured meshes) in around 0.5 seconds, outperforming other open image-to-3D models such as OpenLRM.

TripsoSR: 3D object generation from a single image in <1s
TripsoSR: 3D object generation from a single image in <1s

Why does it matter?

TripoSR caters to the growing demands of various industries, including entertainment, gaming, industrial design, and architecture. The availability of the model weights and source code for download further promotes commercialized, personal, and research use, making it a valuable asset for developers, designers, and creators.

Source

Cloudflare’s Firewall for AI protects LLMs from abuses

Cloudflare has released a Firewall for AI, a protection layer that you can deploy in front of Large Language Models (LLMs) to identify abuses before they reach the models. While the traditional web and API vulnerabilities also apply to the LLM world, Firewall for AI is an advanced-level Web Application Firewall (WAF) designed explicitly for LLM protection and placed in front of applications to detect vulnerabilities and provide visibility to model owners.

Cloudflare Firewall for AI is deployed like a traditional WAF, where every API request with an LLM prompt is scanned for patterns and signatures of possible attacks. You can deploy it in front of models hosted on the Cloudflare Workers AI platform or any other third-party infrastructure. You can use it alongside Cloudflare AI Gateway and control/set up a Firewall for AI using the WAF control plane.

Cloudflare's Firewall for AI protects LLMs from abuses
Cloudflare’s Firewall for AI protects LLMs from abuses

Why does it matter?

As the use of LLMs becomes more widespread, there is an increased risk of vulnerabilities and attacks that malicious actors can exploit. Cloudflare is one of the first security providers to launch tools to secure AI applications. Using a Firewall for AI, you can control what prompts and requests reach their language models, reducing the risk of abuses and data exfiltration. It also aims to provide early detection and protection for both users and LLM models, enhancing the security of AI applications.

Source

🤔 Microsoft compares New York Times to ’80s movie studios trying to ban VCRs

  • Microsoft filed a motion to dismiss the New York Times’ copyright infringement lawsuit against OpenAI, comparing the newspaper’s stance to 1980s movie studios’ attempts to block VCRs, arguing that generative AI, like the VCR, does not hinder the original content’s market.
  • The company, as OpenAI’s largest supporter, asserts that copyright law does not obstruct ChatGPT’s development because the training content does not substantially affect the market for the original content.
  • Microsoft and OpenAI contend that ChatGPT does not replicate or substitute for New York Times content, emphasizing that the AI’s training on such articles does not significantly contribute to its development.
  • Source

🥴 Google co-founder says company ‘definitely messed up’

  • Sergey Brin admitted Google “definitely messed up” with the Gemini AI’s image generation, highlighting issues like historically inaccurate images and the need for more thorough testing.
  • Brin, a core contributor to Gemini, came out of retirement due to the exciting trajectory of AI, amidst the backdrop of Google’s “code red” in response to OpenAI’s ChatGPT.
  • Criticism of Gemini’s biases and errors, including its portrayal of people of color and responses in written form, led to Brin addressing concerns over the AI’s unintended left-leaning output.
  • Source

A Daily Chronicle of AI Innovations in March 2024 – Day 4: AI Daily News – March 04th, 2024

👀 Google’s ScreenAI can ‘see’ graphics like humans do
🐛 How AI ‘worms’ pose security threats in connected systems
🧠 New benchmarking method challenges LLMs’ reasoning abilities

💊 AI may enable personalized prostate cancer treatment

🎥 Vimeo debuts AI-powered video hub for business collaboration

📱 Motorola revving up for AI-powered Moto X50 Ultra launch

📂 Copilot will soon fetch and parse your OneDrive files

⚡ Huawei’s new AI chip threatens Nvidia’s dominance in China

OpenAI adds ‘Read Aloud’ voiceover to ChatGPT

https://youtu.be/ZJvTv7zVX0s?si=yejANUAUtUwyXEH8

OpenAI rolled out a new “Read Aloud” feature for ChatGPT as rivals like Anthropic and Google release more capable language models. (Source)

The Voiceover Update

  • ChatGPT can now narrate responses out loud on mobile apps and web.

  • Activated by tapping the response or clicking the microphone icon.

  • Update comes as Anthropic unveils their newest Claude 3 model.

  • Timing seems reactive amid intense competition over advanced AI. OpenAI also facing lawsuit from Elon Musk over alleged betrayal.

Anthropic launches Claude 3, claiming to outperform GPT-4 across the board

https://youtu.be/Re0WgPNiLo4?si=DwfGraTvhVo8kjuK

Here’s the announcement from Anthropic and their benchmark results:
https://twitter.com/AnthropicAI/status/1764653830468428150

Anthropic launches Claude 3, claiming to outperform GPT-4 across the board
Anthropic launches Claude 3, claiming to outperform GPT-4 across the board

Google’s ScreenAI can ‘see’ graphics like humans do

Google Research has introduced ScreenAI, a Vision-Language Model that can perform question-answering on digital graphical content like infographics, illustrations, and maps while also annotating, summarizing, and navigating UIs. The model combines computer vision (PaLI architecture) with text representations of images to handle these multimodal tasks.

Despite having just 4.6 billion parameters, ScreenAI achieves new state-of-the-art results on UI- and infographics-based tasks and new best-in-class performance on others, compared to models of similar size.

Google’s ScreenAI can ‘see’ graphics like humans do
Google’s ScreenAI can ‘see’ graphics like humans do

While ScreenAI is best-in-class on some tasks, further research is needed to match models like GPT-4 and Gemini, which are significantly larger. Google Research has released a dataset with ScreenAI’s unified representation and two other datasets to help the community experiment with more comprehensive benchmarking on screen-related tasks.

Why does this matter?

ScreenAI’s breakthrough in unified visual and language understanding bridges the disconnect between how humans and machines interpret ideas across text, images, charts, etc. Companies can now leverage these multimodal capabilities to build assistants that summarize reports packed with graphics, analysts that generate insights from dashboard visualizations, and agents that manipulate UIs to control workflows.

Source

How AI ‘worms’ pose security threats in connected systems

Security researchers have created an AI “worm” called Morris II to showcase vulnerabilities in AI ecosystems where different AI agents are linked together to complete tasks autonomously.

The researchers tested the worm in a simulated email system using ChatGPT, Gemini, and other popular AI tools. The worm can exploit these AI systems to steal confidential data from emails or forward spam/propaganda without human approval. It works by injecting adversarial prompts that make the AI systems behave maliciously.

While this attack was simulated, the research highlights risks if AI agents are given too much unchecked freedom to operate.

Why does it matter?

This AI “worm” attack reveals that generative models like ChatGPT have reached capabilities that require heightened security to prevent misuse. Researchers and developers must prioritize safety by baking in controls and risk monitoring before commercial release. Without industry-wide commitments to responsible AI, regulation may be needed to enforce acceptable safeguards across critical domains as systems gain more autonomy.

Source

New benchmarking method challenges LLMs’ reasoning abilities

Researchers at Consequent AI have identified a “reasoning gap” in large language models like GPT-3.5 and GPT-4. They introduced a new benchmarking approach called “functional variants,” which tests a model’s ability to reason instead of just memorize. This involves translating reasoning tasks like math problems into code that can generate unique questions requiring the same logic to solve.

New benchmarking method challenges LLMs’ reasoning abilities
New benchmarking method challenges LLMs’ reasoning abilities

When evaluating several state-of-the-art models, the researchers found a significant gap between performance on known problems from benchmarks versus new problems the models had to reason through. The gap was 58-80%, indicating the models do not truly understand complex problems but likely just store training examples. The models performed better on simpler math but still demonstrated limitations in reasoning ability.

Why does this matter?

This research reveals that reasoning still eludes our most advanced AIs. We risk being misled by claims of progress made by the Big Tech if their benchmarks reward superficial tricks over actual critical thinking. Moving forward, model creators will have to prioritize generalization and logic over memorization if they want to make meaningful progress towards general intelligence.

Source

What Else Is Happening in AI on March 04th, 2024❗

💊 AI may enable personalized prostate cancer treatment

Researchers used AI to analyze prostate cancer DNA and found two distinct subtypes called “evotypes.” Identifying these subtypes could allow for better prediction of prognosis and personalized treatments. (Link)

🎥 Vimeo debuts AI-powered video hub for business collaboration

Vimeo has launched a new product called Vimeo Central, an AI-powered video hub to help companies improve internal video communications, collaboration, and analytics. Key capabilities include a centralized video library, AI-generated video summaries and highlights, enhanced screen recording and video editing tools, and robust analytics. (Link)

📱 Motorola revving up for AI-powered Moto X50 Ultra launch

Motorola is building hype for its upcoming Moto X50 Ultra phone with a Formula 1-themed teaser video highlighting the device’s powerful AI capabilities. The phone will initially launch in China on April 21 before potentially getting a global release under the Motorola Edge branding. (Link)

📂 Copilot will soon fetch and parse your OneDrive files

Microsoft is soon to launch Copilot for OneDrive, an AI assistant that will summarize documents, extract information, answer questions, and follow commands related to files stored in OneDrive. Copilot can generate outlines, tables, and lists based on documents, as well as tailored summaries and responses. (Link)

⚡ Huawei’s new AI chip threatens Nvidia’s dominance in China

Huawei has developed a new AI chip, the Ascend 910B, which matches the performance of Nvidia’s A100 GPU based on assessments by SemiAnalysis. The Ascend 910B is already being used by major Chinese companies like Baidu and iFlytek and could take market share from Nvidia in China due to US export restrictions on Nvidia’s latest AI chips. (Link)

1-bit LLMs explained

Check out this new tutorial that summarizes the revolutionary paper “The Era of 1-bit LLMs” introducing BitNet b1.58 model and explain what are 1-bit LLMs and how they are useful.

A Daily Chronicle of AI Innovations in March 2024 – Day 2: AI Daily News – March 02nd, 2024

A Daily Chronicle of AI Innovations in March 2024 – Day 1: AI Daily News – March 01st, 2024

🪄Sora showcases jaw-dropping geometric consistency
🧑‍✈️Microsoft introduces Copilot for finance in Microsoft 365
🤖OpenAI and Figure team up to develop AI for robots

Elon Sues OpenAI for “breach of contract”

Elon Musk filed suit against OpenAI and CEO Sam Altman, alleging they have breached the artificial-intelligence startup’s founding agreement by putting profit ahead of benefiting humanity.

The 52-year-old billionaire, who helped fund OpenAI in its early days, said the company’s close relationship with Microsoft has undermined its original mission of creating open-source technology that wouldn’t be subject to corporate priorities. Musk, who is also CEO of Tesla has been among the most outspoken about the dangers of AI and artificial general intelligence, or AGI.

“To this day, OpenAI Inc.’s website continues to profess that its charter is to ensure that AGI “benefits all of humanity.” In reality, however, OpenAI has been transformed into a closed-source de facto subsidiary of the largest technology company in the world: Microsoft,” the lawsuit says.

ELON MUSK vs. SAMUEL ALTMAN, GREGORY BROCKMAN, OPENAI, INC.
Elon Sues OpenAI for “breach of contract”

Sora showcases jaw-dropping geometric consistency

Sora from OpenAI has been remarkable in video generation compared to other leading models like Pika and Gen2. In a recent benchmarking test conducted by ByteDanc.Inc in collaboration with Wuhan and Nankai University, Sora showcased video generation with high geometric consistency.

AI Innovations in March 2024: Sora showcases jaw-dropping geometric consistency
Sora showcases jaw-dropping geometric consistency

The benchmark test assesses the quality of generated videos based on how it adhere to the principles of physics in real-world scenarios. Researchers used an approach where generated videos are transformed into 3D models. Further, a team of researchers used the fidelity of geometric constraints to measure the extent to which generated videos conform to physics principles in the real world.

Why does it matter?

Sora’s remarkable performance in generating geometrically consistent videos can greatly boost several use cases for construction engineers and architects. Further, the new benchmarking will allow researchers to measure newly developed models to understand how accurately their creations conform to the principles of physics in real-world scenarios.

Source

Microsoft introduces Copilot for finance in Microsoft 365

Microsoft has launched Copilot for Finance, a new addition to its Copilot series that recommends AI-powered productivity enhancements. It aims to transform how finance teams approach their daily work with intelligent workflow automation, recommendations, and guided actions. This Copilot aims to simplify data-driven decision-making, helping finance professionals have more free time by automating manual tasks like Excel and Outlook.

Copilot for Finance simplifies complex variance analysis in Excel, account reconciliations, and customer account summaries in Outlook. Dentsu, Northern Trust, Schneider Electric, and Visa plan to use it alongside Copilot for Sales and Service to increase productivity, reduce case handling times, and gain better decision-making insights.

Why does it matter?

Introducing Microsoft Copilot for finance will help businesses focus on strategic involvement from professionals otherwise busy with manual tasks like data entry, workflow management, and more. This is a great opportunity for several organizations to automate tasks like analysis of anomalies, improve analytic efficiency, and expedite financial transactions.

Source

OpenAI and Figure team up to develop AI for robots 

Figure has raised $675 million in series B funding with investments from OpenAI, Microsoft, and NVIDIA. It is an AI robotics company developing humanoid robots for general-purpose usage. The collaboration agreement between OpenAI and Figure aims to develop advanced humanoid robots that will leverage the generative AI models at its core.

This collaboration will also help accelerate the development of smart humanoid robots capable of understanding tasks like humans. With its deep understanding of robotics, Figure is set to bring efficient robots for general-purpose enhancing automation.

Why does it matter?

Open AI and Figure will transform robot operations, adding generative AI capabilities. This collaboration will encourage the integration of generative AI capabilities across robotics development. Right from industrial robots to general purpose and military applications, generative AI can be the new superpower for robotic development.

Source

🔍 Google now wants to limit the AI-powered search spam it helped create

  • Google announced it will tackle AI-generated content aiming to manipulate search rankings through algorithmic enhancements, affecting automated content creation the most.
  • These algorithm changes are intended to discern and reduce low-quality and unhelpful webpages, aiming to improve the overall quality of search results.
  • The crackdown also targets misuse of high-reputation websites and the exploitation of expired domains for promoting substandard content.
  • Source

What Else Is Happening in AI in March 2024❗

🤝Stack Overflow partners with Google Cloud to power AI 

Stack Overflow and Google Cloud are partnering to integrate OverflowAPI into Google Cloud’s AI tools. This will give developers accessing the Google Cloud console access to Stack Overflow’s vast knowledge base of over 58 million questions and answers. The partnership aims to enable AI systems to provide more insightful and helpful responses to users by learning from the real-world experiences of programmers. (Link)

💻Microsoft unites rival GPU makers for one upscaling API

Microsoft is working with top graphics hardware makers to introduce “DirectSR”, a new API that simplifies the integration of super-resolution upscaling into games. DirectSR will allow game developers to easily access Nvidia’s DLSS, AMD’s FSR, and Intel’s XeSS with a single code path. Microsoft will preview the API in its Agility SDK soon and demonstrate it live with AMD and Nvidia reps on March 21st. (Link)

📈Google supercharges data platforms with AI for deeper insights

Google is expanding its AI capabilities across data and analytics services, including BigQuery and Cloud Databases. Vector search support is available across all databases, and BigQuery has the advanced Gemini Pro model for unstructured data analysis. Users can combine insights from images, video, audio, and text with structured data in a single analytics workflow. (Link)

🔍 Brave’s privacy-first AI-powered assistant is now available on Android 

Brave’s AI-powered assistant, Leo, is now available on Android, bringing helpful features like summarization, transcription, and translation while prioritizing user privacy. Leo processes user inputs locally on the device without retaining or using data to train itself, aligning with Brave’s commitment to privacy-focused services. Users can simplify tasks with Leo without compromising on security. (Link)

Elsewhere in AI anxiety:

February 2024 AI Recap

February 2024 AI Recap
February 2024 AI Recap

February 2024 – Week 4 Recap

  1. Mistral introduced a new model Mistral Large. It reaches top-tier reasoning capabilities, is multi-lingual by design, has native function calling capacities and has 32K tokens context window. The pre-trained model has 81.2% accuracy on MMLU. Alongside Mistral Large, Mistral Small, a model optimized for latency and cost has been released. Mistral Small outperforms Mixtral 8x7B and has lower latency. Mistral also launched a ChatGPT like new conversational assistant, le Chat Mistral [Details].
  2. Alibaba Group introduced EMO, an expressive audio-driven portrait-video generation framework. Input a single reference image and the vocal audio, e.g. talking and singing, it can generate vocal avatar videos with expressive facial expressions, and various head poses [Details].
  3. Ideogram introduced Ideogram 1.0, a text-to-image model trained from scratch for state-of-the-art text rendering, photorealism, prompt adherence, and a feature called Magic Prompt to help with prompting. Ideogram 1.0 is now available to all users on ideogram.ai [Details].

    Ideogram introduced Ideogram 1.0
    Ideogram introduced Ideogram 1.0

  4. Google DeepMind introduced Genie (generative interactive environments), a foundation world model trained exclusively from Internet videos that can generate interactive, playable environments from a single image prompt  [Details].
  5. Pika Labs launched Lip Sync feature, powered by audio from Eleven Labs, for its AI generated videos enabling users to make the characters talk with realistic mouth movements [Video].
  6.  UC Berkeley introduced Berkeley Function Calling Leaderboard (BFCL) to evaluate the function calling capability of different LLMs. Gorilla Open Functions v2, an open-source model that can help users with building AI applications with function calling and interacting with json compatible output has also been released [Details].
  7. Qualcomm launched AI Hub, a curated library of 80+ optimized AI models for superior on-device AI performance across Qualcomm and Snapdragon platforms [Details].
  8. BigCode released StarCoder2, a family of open LLMs for code and comes in 3 different sizes with 3B, 7B and 15B parameters. StarCoder2-15B is trained on over 4 trillion tokens and 600+ programming languages from The Stack v2 dataset [Details].
  9. Researchers released FuseChat-7B-VaRM, which is the fusion of three prominent chat LLMs with diverse architectures and scales, namely NH2-Mixtral-8x7B, NH2-Solar-10.7B, and OpenChat-3.5-7B, surpassing GPT-3.5 (March), Claude-2.1, and approaching Mixtral-8x7B-Instruct [Details].
  10. The Swedish fintech Klarna’s AI assistant handles two-thirds of all customer service chats, some 2.3 million conversations so far, equivalent to the work of 700 people [Details].
  11. Lightricks introduces LTX Studio, an AI-powered film making platform, now open for waitlist sign-ups, aimed at assisting creators in story visualization [Details].
  12. Morph partners with Stability AI to launch Morph Studio, a platform to make films using Stability AI–generated clips [Details].
  13. JFrog‘s security team found that roughly a 100 models hosted on the Hugging Face platform feature malicious functionality [Details].
  14. Playground released Playground v2.5, an open-source text-to-image generative model, with a focus on enhanced color and contrast, improved generation for multi-aspect ratios, and improved human-centric fine detail [Details].
  15. Together AI and the Arc Institute released Evo, a long-context biological foundation model based on the StripedHyena architecture that generalizes across DNA, RNA, and proteins.. Evo is capable of both prediction tasks and generative design, from molecular to whole genome scale (over 650k tokens in length) [Details].
  16. Adobe previews a new generative AI music generation and editing tool, Project Music GenAI Control, that allows creators to generate music from text prompts, and then have fine-grained control to edit that audio for their precise needs [Details | video].
  17. Microsoft introduces Copilot for Finance, an AI chatbot for finance workers in Excel and Outlook [Details].
  18. The Intercept, Raw Story, and AlterNet sue OpenAI and Microsoft, claiming OpenAI and Microsoft intentionally removed important copyright information from training data [Details].
  19. Huawei spin-off Honor shows off tech to control a car with your eyes and chatbot based on Meta’s AI [Details].
  20. Tumblr and WordPress.com are preparing to sell user data to Midjourney and OpenAI [Details]

February 2024 – Week 3 Recap

  1. Meta AI introduces V-JEPA (Video Joint Embedding Predictive Architecture), a method for teaching machines to understand and model the physical world by watching videos. Meta AI releases a collection of V-JEPA vision models trained with a feature prediction objective using self-supervised learning. The models are able to understand and predict what is going on in a video, even with limited information [Details | GitHub].
  2. Open AI introduces Sora, a text-to-video model that can create videos of up to 60 seconds featuring highly detailed scenes, complex camera motion, and multiple characters with vibrant emotions [Details + sample videos Report].
  3. Google announces their next-generation model, Gemini 1.5, that uses a new Mixture-of-Experts (MoE) architecture. The first Gemini 1.5 model being released for early testing is Gemini 1.5 Pro with a context window of up to 1 million tokens, which is the longest context window of any large-scale foundation model yet. 1.5 Pro can perform sophisticated understanding and reasoning tasks for different modalities, including video and it performs at a similar level to 1.0 Ultra [Details |Tech Report].
  4. Reka introduced Reka Flash, a new 21B multimodal and multilingual model trained entirely from scratch that is competitive with Gemini Pro & GPT 3.5 on key language & vision benchmarks. Reka also present a compact variant Reka Edge , a smaller and more efficient model (7B) suitable for local and on-device deployment. Both models are in public beta and available in Reka Playground [Details].
  5. Cohere For AI released Aya, a new open-source, massively multilingual LLM & dataset to help support under-represented languages. Aya outperforms existing open-source models and covers 101 different languages – more than double covered by previous models [Details].
  6. BAAI released Bunny, a family of lightweight but powerful multimodal models. Bunny-3B model built upon SigLIP and Phi-2 outperforms the state-of-the-art MLLMs, not only in comparison with models of similar size but also against larger MLLMs (7B), and even achieves performance on par with LLaVA-13B [Details].
  7. Amazon introduced a text-to-speech (TTS) model called BASE TTS (Big Adaptive Streamable TTS with Emergent abilities). BASE TTS is the largest TTS model to-date, trained on 100K hours of public domain speech data and exhibits “emergent” qualities improving its ability to speak even complex sentences naturally [Details | Paper].
  8. Stability AI released Stable Cascade in research preview, a new text to image model that is exceptionally easy to train and finetune on consumer hardware due to its three-stage architecture. Stable Cascade can also generate image variations and image-to-image generations. In addition to providing checkpoints and inference scripts, Stability AI has also released scripts for finetuning, ControlNet, and LoRA training [Details].
  9. Researchers from UC berkeley released Large World Model (LWM), an open-source general-purpose large-context multimodal autoregressive model, trained from LLaMA-2, that can perform language, image, and video understanding and generation. LWM answers questions about 1 hour long YouTube video even if GPT-4V and Gemini Pro both fail and can retriev facts across 1M context with high accuracy [Details].
  10. GitHub opens applications for the next cohort of GitHub Accelerator program with a focus on funding the people and projects that are building AI-based solutions under an open source license [Details].
  11. NVIDIA released Chat with RTX, a locally running (Windows PCs with specific NVIDIA GPUs) AI assistant that integrates with your file system and lets you chat with your notes, documents, and videos using open source models [Details].
  12. Open AI is testing memory with ChatGPT, enabling it to remember things you discuss across all chats. ChatGPT’s memories evolve with your interactions and aren’t linked to specific conversations. It is being rolled out to a small portion of ChatGPT free and Plus users this week [Details].
  13. BCG X released of AgentKit, a LangChain-based starter kit (NextJS, FastAPI) to build constrained agent applications [Details | GitHub].
  14. Elevenalabs’ Speech to Speech feature, launched in November, for voice transformation with control over emotions and delivery, is now multilingual and available in 29 languages [Link]
  15. Apple introduced Keyframer, an LLM-powered animation prototyping tool that can generate animations from static images (SVGs). Users can iterate on their design by adding prompts and editing LLM-generated CSS animation code or properties [Paper].
  16. Eleven Labs launched a payout program for voice actors to earn rewards every time their voice clone is used [Details].
  17. Azure OpenAI Service announced Assistants API, new models for finetuning, new text-to-speech model and new generation of embeddings models with lower pricing [Details].
  18. Brilliant Labs, the developer of AI glasses, launched Frame, the world’s first glasses featuring an integrated AI assistant, Noa. Powered by an integrated multimodal generative AI system capable of running GPT4, Stability AI, and the Whisper AI model simultaneously, Noa performs real-world visual processing, novel image generation, and real-time speech recognition and translation. [Details].
  19. Nous Research released Nous Hermes 2 Llama-2 70B model trained on the Nous Hermes 2 dataset, with over 1,000,000 entries of primarily synthetic data [Details].
  20. Open AI in partnership with Microsoft Threat Intelligence, have disrupted five state-affiliated actors that sought to use AI services in support of malicious cyber activities [Details]
  21. Perplexity partners with Vercel, opening AI search to developer apps [Details].
  22. Researchers show that LLM agents can autonomously hack websites.

February 2024 – Week 2 Recap:

  1. Google launches Ultra 1.0, its largest and most capable AI model, in its ChatGPT-like assistant which has now been rebranded as Gemini (earlier called Bard). Gemini Advanced is available, in 150 countries, as a premium plan for $19.99/month, starting with a two-month trial at no cost. Google is also rolling out Android and iOS apps for Gemini [Details].
  2. Alibaba Group released Qwen1.5 series, open-sourcing models of 6 sizes: 0.5B, 1.8B, 4B, 7B, 14B, and 72B. Qwen1.5-72B outperforms Llama2-70B across all benchmarks. The Qwen1.5 series is available on Ollama and LMStudio. Additionally, API on together.ai [Details | Hugging Face].
  3. NVIDIA released Canary 1B, a multilingual model for speech-to-text recognition and translation. Canary transcribes speech in English, Spanish, German, and French and also generates text with punctuation and capitalization. It supports bi-directional translation, between English and three other supported languages. Canary outperforms similarly-sized Whisper-large-v3, and SeamlessM4T-Medium-v1 on both transcription and translation tasks and achieves the first place on HuggingFace Open ASR leaderboard with an average word error rate of 6.67%, outperforming all other open source models [Details].
  4. Researchers released Lag-Llama, the first open-source foundation model for time series forecasting [Details].
  5. LAION released BUD-E, an open-source conversational and empathic AI Voice Assistant that uses natural voices, empathy & emotional intelligence and can handle multi-speaker conversations [Details].
  6. MetaVoice released MetaVoice-1B, a 1.2B parameter base model trained on 100K hours of speech, for TTS (text-to-speech). It supports emotional speech in English and voice cloning. MetaVoice-1B has been released under the Apache 2.0 license [Details].
  7. Bria AI released RMBG v1.4, an an open-source background removal model trained on fully licensed images [Details].
  8. Researchers introduce InteractiveVideo, a user-centric framework for video generation that is designed for dynamic interaction, allowing users to instruct the generative model during the generation process [Details |GitHub ].
  9. Microsoft announced a redesigned look for its Copilot AI search and chatbot experience on the web (formerly known as Bing Chat), new built-in AI image creation and editing functionality, and Deucalion, a fine tuned model that makes Balanced mode for Copilot richer and faster [Details].
  10. Roblox introduced AI-powered real-time chat translations in 16 languages [Details].
  11. Hugging Face launched Assistants feature on HuggingChat. Assistants are custom chatbots similar to OpenAI’s GPTs that can be built for free using open source LLMs like Mistral, Llama and others [Link].
  12. DeepSeek AI released DeepSeekMath 7B model, a 7B open-source model that approaches the mathematical reasoning capability of GPT-4. DeepSeekMath-Base is initialized with DeepSeek-Coder-Base-v1.5 7B [Details].
  13. Microsoft is launching several collaborations with news organizations to adopt generative AI [Details].
  14. LG Electronics signed a partnership with Korean generative AI startup Upstage to develop small language models (SLMs) for LG’s on-device AI features and AI services on LG notebooks [Details].
  15. Stability AI released SVD 1.1, an updated model of Stable Video Diffusion model, optimized to generate short AI videos with better motion and more consistency [Details | Hugging Face] .
  16. OpenAI and Meta announced to label AI generated images [Details].
  17. Google saves your conversations with Gemini for years by default [Details].

February 2024 – Week 1 Recap:

  1. Amazon presents Diffuse to Choose, a diffusion-based image-conditioned inpainting model that allows users to virtually place any e-commerce item in any setting, ensuring detailed, semantically coherent blending with realistic lighting and shadows. Code and demo will be released soon [Details].
  2. OpenAI announced two new embedding models, new GPT-4 Turbo and moderation models, new API usage management tools, and lower pricing on GPT-3.5 Turbo. The updated GPT-4 Turbo preview model reduces cases of “laziness” where the model doesn’t complete a task. The new embedding models include a smaller and highly efficient text-embedding-3-small model, and a larger and more powerful text-embedding-3-large model. [Details].
  3. Hugging Face and Google partner to support developers building AI applications [Details].
  4. Adept introduced Adept Fuyu-Heavy, a new multimodal model designed specifically for digital agents. Fuyu-Heavy scores higher on the MMMU benchmark than Gemini Pro [Details].
  5. Fireworks.ai has open-sourced FireLLaVA, a LLaVA multi-modality model trained on OSS LLM generated instruction following data, with a commercially permissive license. Firewroks.ai is also providing both the completions API and chat completions API to devlopers [Details].
  6. 01.AI released Yi Vision Language (Yi-VL) model, an open-source, multimodal version of the Yi Large Language Model (LLM) series, enabling content comprehension, recognition, and multi-round conversations about images. Yi-VL adopts the LLaVA architecture and is free for commercial use. Yi-VL-34B is the first open-source 34B vision language model worldwide [Details].
  7. Tencent AI Lab introduced WebVoyager, an innovative Large Multimodal Model (LMM) powered web agent that can complete user instructions end-to-end by interacting with real-world websites [Paper].
  8. Prophetic introduced MORPHEUS-1, a multi-modal generative ultrasonic transformer model designed to induce and stabilize lucid dreams from brain states. Instead of generating words, Morpheus-1 generates ultrasonic holograms for neurostimulation to bring one to a lucid state [Details].
  9. Google Research presented Lumiere – a space-time video diffusion model for text-to-video, image-to-video, stylized generation, inpainting and cinemagraphs [Details].
  10. TikTok released Depth Anything, an image-based depth estimation method trained on 1.5M labeled images and 62M+ unlabeled images jointly [Details].
  11. Nightshade, the free tool that ‘poisons’ AI models, is now available for artists to use [Details].
  12. Stability AI released Stable LM 2 1.6B, 1.6 billion parameter small language model trained on multilingual data in English, Spanish, German, Italian, French, Portuguese, and Dutch. Stable LM 2 1.6B can be used now both commercially and non-commercially with a Stability AI Membership [Details].
  13. Etsy launched ‘Gift Mode,’ an AI-powered feature designed to match users with tailored gift ideas based on specific preferences [Details].
  14. Google DeepMind presented AutoRT, a framework that uses foundation models to scale up the deployment of operational robots in completely unseen scenarios with minimal human supervision. In AutoRT, a VLM describes the scene, an LLM generates robot goals and filters for affordance and safety, then routes execution to policies [Details].
  15. Google Chrome gains AI features, including a writing helper, theme creator, and tab organizer [Details].
  16. Tencent AI Lab released VideoCrafter2 for high quality text-to-video generation, featuring major improvements in visual quality, motion and concept Composition compared to VideoCrafter1 [Details | Demo]
  17. Google opens beta access to the conversational experience, a new chat-based feature in Google Ads, for English language advertisers in the U.S. & U.K. It will let advertisers create optimized Search campaigns from their website URL by generating relevant ad content, including creatives and keywords [Details].

A Daily Chronicle of AI Innovations in February 2024

  • AI: The Ultimate Sherlocking?
    by /u/mintone (Artificial Intelligence) on July 26, 2024 at 12:16 pm

    submitted by /u/mintone [link] [comments]

  • Speech-to-Text Solution for Multilingual Sentences / Mixed-language speech
    by /u/simbaninja33 (Artificial Intelligence Gateway) on July 26, 2024 at 11:54 am

    I am looking for a speech-to-text solution, either paid or open-source, that can accurately transcribe speech containing a mix of two languages within the same sentence. I have explored options like Microsoft Azure, Google Cloud, and OpenAI, but haven't found a satisfactory solution yet. For example, I need the solution to handle sentences like: "I have tried the restaurant yesterday, it is muy muy bueno, they serve some of the pizza, que haria mi abuela super celoza de la receta." "I went to the store y compré un poco de pan because we were running low." I have already tried Microsoft Azure, which can handle multiple languages, but only when they are not mixed within the same sentence (as mentioned in their documentation). Google Cloud's speech-to-text fails to accurately transcribe mixed-language speech, and OpenAI doesn't seem to offer this functionality. I am open to both continuous real-time speech recognition and file-based recognition. For real-time applications, I am also willing to consider workarounds, such as implementing a "button" that can be clicked to quickly switch between the main language and the second language. If anyone has experience with a solution that can handle this type of mixed-language speech recognition, I would greatly appreciate any suggestions or recommendations. Thank you in advance for your help! submitted by /u/simbaninja33 [link] [comments]

  • Any open source AI model with web search abilities?
    by /u/david8840 (Artificial Intelligence Gateway) on July 26, 2024 at 11:45 am

    Is there any open source AI model with web search abilities? I want to be able to ask it questions which require real time internet searching, for example "What is the weather like now in NY?" submitted by /u/david8840 [link] [comments]

  • Which companies are leading the way in AI detection? (for audio/video deepfakes, etc.?)
    by /u/ProfessionalHat3555 (Artificial Intelligence Gateway) on July 26, 2024 at 11:21 am

    So I was listening to the most recent Bill Simmons pod w/ Derek Thompson where they discuss conspiracy theories and AI shit-detection (40:00-48:00 if you're curious)... 1ST Q: what companies are you aware of that are already working on AI detection? 2ND Q: where do you think the AI detection slice of the market is going? Will there be consumer-grade products that we can use to run, say, a political video through a detection software & get a % of realness rating on it? Will these tools ONLY be available to big conglomerates who become the purveyors of truth? 3RD Q: If we're UNABLE to do this at-scale yet, what would need to happen tech-wise for AI detection to become more accessible to more people? (disclaimer: I'm not a dev) submitted by /u/ProfessionalHat3555 [link] [comments]

  • AI can't take people's jobs if there's no people.
    by /u/baalzimon (Artificial Intelligence Gateway) on July 26, 2024 at 10:53 am

    Looks more and more likely that human populations will decline in the future. Maybe the workforce will just be AI robots rather than young people. PEW: The Experiences of U.S. Adults Who Don’t Have Children 57% of adults under 50 who say they’re unlikely to ever have kids say a major reason is they just don’t want to; 31% of those ages 50 and older without kids cite this as a reason they never had them https://www.pewresearch.org/social-trends/2024/07/25/the-experiences-of-u-s-adults-who-dont-have-children/ submitted by /u/baalzimon [link] [comments]

  • UK School Under Fire for Unlawful Facial-Recognition Use
    by /u/Think_Cat1101 (Artificial Intelligence Gateway) on July 26, 2024 at 10:43 am

    https://www.msn.com/en-us/news/technology/uk-school-under-fire-for-unlawful-facial-recognition-use/ar-BB1qEmeX?cvid=6dfe65854c6e4c2ad473b0e649e795b2&ei=10 submitted by /u/Think_Cat1101 [link] [comments]

  • OpenAI reveals 'SearchGPT'
    by /u/Mindful-AI (Artificial Intelligence Gateway) on July 26, 2024 at 10:41 am

    submitted by /u/Mindful-AI [link] [comments]

  • Amazon’s AI Chip Revolution: How They’re Ditching Nvidia’s High Prices and Speeding Ahead
    by /u/alyis4u (Artificial Intelligence Gateway) on July 26, 2024 at 9:23 am

    Six engineers tested a brand-new, secret server design on a Friday afternoon in Amazon.com’s chip lab in Austin, Texas. Amazon executive Rami Sinno said on Friday during a visit to the lab that the server was full of Amazon’s AI chips, which compete with Nvidia’s chips and are the market leader.https://theaiwired.com/amazons-ai-chip-revolution-how-theyre-ditching-nvidias-high-prices-and-speeding-ahead/ submitted by /u/alyis4u [link] [comments]

  • OpenAI's SearchGPT Is Coming For Google Search; Here Are The Features That Will Reportedly Make It Better
    by /u/vinaylovestotravel (Artificial Intelligence Gateway) on July 26, 2024 at 9:00 am

    Dubbed "SearchGPT," the tool will offer "fast and timely answers with clear and relevant sources" by referencing content from websites and news publishers, including OpenAI content partners such as News Corp (The Post's parent company) and The Atlantic. Read more: https://www.ibtimes.co.uk/openais-searchgpt-coming-google-search-here-are-features-that-will-reportedly-make-it-better-1725770 submitted by /u/vinaylovestotravel [link] [comments]

  • Deleting chats from Blackbox AI?
    by /u/Intelligent-Fig-7791 (Artificial Intelligence Gateway) on July 26, 2024 at 7:40 am

    How on earth do you delete chats from blackbox.ai ? it seems like all chats are public by default submitted by /u/Intelligent-Fig-7791 [link] [comments]

Ace the 2023 AWS Solutions Architect Associate SAA-C03 Exam with Confidence Pass the 2023 AWS Certified Machine Learning Specialty MLS-C01 Exam with Flying Colors

List of Freely available programming books - What is the single most influential book every Programmers should read



#BlackOwned #BlackEntrepreneurs #BlackBuniness #AWSCertified #AWSCloudPractitioner #AWSCertification #AWSCLFC02 #CloudComputing #AWSStudyGuide #AWSTraining #AWSCareer #AWSExamPrep #AWSCommunity #AWSEducation #AWSBasics #AWSCertified #AWSMachineLearning #AWSCertification #AWSSpecialty #MachineLearning #AWSStudyGuide #CloudComputing #DataScience #AWSCertified #AWSSolutionsArchitect #AWSArchitectAssociate #AWSCertification #AWSStudyGuide #CloudComputing #AWSArchitecture #AWSTraining #AWSCareer #AWSExamPrep #AWSCommunity #AWSEducation #AzureFundamentals #AZ900 #MicrosoftAzure #ITCertification #CertificationPrep #StudyMaterials #TechLearning #MicrosoftCertified #AzureCertification #TechBooks

Top 1000 Canada Quiz and trivia: CANADA CITIZENSHIP TEST- HISTORY - GEOGRAPHY - GOVERNMENT- CULTURE - PEOPLE - LANGUAGES - TRAVEL - WILDLIFE - HOCKEY - TOURISM - SCENERIES - ARTS - DATA VISUALIZATION
zCanadian Quiz and Trivia, Canadian History, Citizenship Test, Geography, Wildlife, Secenries, Banff, Tourism

Top 1000 Africa Quiz and trivia: HISTORY - GEOGRAPHY - WILDLIFE - CULTURE - PEOPLE - LANGUAGES - TRAVEL - TOURISM - SCENERIES - ARTS - DATA VISUALIZATION
Africa Quiz, Africa Trivia, Quiz, African History, Geography, Wildlife, Culture

Exploring the Pros and Cons of Visiting All Provinces and Territories in Canada.
Exploring the Pros and Cons of Visiting All Provinces and Territories in Canada

Exploring the Advantages and Disadvantages of Visiting All 50 States in the USA
Exploring the Advantages and Disadvantages of Visiting All 50 States in the USA


Health Health, a science-based community to discuss health news and the coronavirus (COVID-19) pandemic

Today I Learned (TIL) You learn something new every day; what did you learn today? Submit interesting and specific facts about something that you just found out here.

Reddit Science This community is a place to share and discuss new scientific research. Read about the latest advances in astronomy, biology, medicine, physics, social science, and more. Find and submit new publications and popular science coverage of current research.

Reddit Sports Sports News and Highlights from the NFL, NBA, NHL, MLB, MLS, and leagues around the world.

Turn your dream into reality with Google Workspace: It’s free for the first 14 days.
Get 20% off Google Google Workspace (Google Meet) Standard Plan with  the following codes:
Get 20% off Google Google Workspace (Google Meet) Standard Plan with  the following codes: 96DRHDRA9J7GTN6 96DRHDRA9J7GTN6
63F733CLLY7R7MM
63F7D7CPD9XXUVT
63FLKQHWV3AEEE6
63JGLWWK36CP7WM
63KKR9EULQRR7VE
63KNY4N7VHCUA9R
63LDXXFYU6VXDG9
63MGNRCKXURAYWC
63NGNDVVXJP4N99
63P4G3ELRPADKQU
With Google Workspace, Get custom email @yourcompany, Work from anywhere; Easily scale up or down
Google gives you the tools you need to run your business like a pro. Set up custom email, share files securely online, video chat from any device, and more.
Google Workspace provides a platform, a common ground, for all our internal teams and operations to collaboratively support our primary business goal, which is to deliver quality information to our readers quickly.
Get 20% off Google Workspace (Google Meet) Business Plan (AMERICAS): M9HNXHX3WC9H7YE
C37HCAQRVR7JTFK
C3AE76E7WATCTL9
C3C3RGUF9VW6LXE
C3D9LD4L736CALC
C3EQXV674DQ6PXP
C3G9M3JEHXM3XC7
C3GGR3H4TRHUD7L
C3LVUVC3LHKUEQK
C3PVGM4CHHPMWLE
C3QHQ763LWGTW4C
Even if you’re small, you want people to see you as a professional business. If you’re still growing, you need the building blocks to get you where you want to be. I’ve learned so much about business through Google Workspace—I can’t imagine working without it.
(Email us for more codes)