

Elevate Your Career with AI & Machine Learning For Dummies PRO and Start mastering the technologies shaping the future—download now and take the next step in your professional journey!
What is OpenAI Q*? A deeper look at the Q* Model as a combination of A* algorithms and Deep Q-learning networks.
Embark on a journey of discovery with our podcast, ‘What is OpenAI Q*? A Deeper Look at the Q* Model’. Dive into the cutting-edge world of AI as we unravel the mysteries of OpenAI’s Q* model, a groundbreaking blend of A* algorithms and Deep Q-learning networks. 🌟🤖
In this detailed exploration, we dissect the components of the Q* model, explaining how A* algorithms’ pathfinding prowess synergizes with the adaptive decision-making capabilities of Deep Q-learning networks. This video is perfect for anyone curious about the intricacies of AI models and their real-world applications.
Understand the significance of this fusion in AI technology and how it’s pushing the boundaries of machine learning, problem-solving, and strategic planning. We also delve into the potential implications of Q* in various sectors, discussing both the exciting possibilities and the ethical considerations.
Join the conversation about the future of AI and share your thoughts on how models like Q* are shaping the landscape. Don’t forget to like, share, and subscribe for more deep dives into the fascinating world of artificial intelligence! #OpenAIQStar #AStarAlgorithms #DeepQLearning #ArtificialIntelligence #MachineLearningInnovation”
🚀 Whether you’re a tech enthusiast, a professional in the field, or simply curious about artificial intelligence, this podcast is your go-to source for all things AI. Subscribe for weekly updates and deep dives into artificial intelligence innovations.
✅ Don’t forget to Like, Comment, and Share this video to support our content.
📌 Check out our playlist for more AI insights
📖 Read along with the podcast:
Welcome to AI Unraveled, the podcast that demystifies frequently asked questions on artificial intelligence and keeps you up to date with the latest AI trends. Join us as we delve into groundbreaking research, innovative applications, and emerging technologies that are pushing the boundaries of AI. From the latest trends in ChatGPT and the recent merger of Google Brain and DeepMind, to the exciting developments in generative AI, we’ve got you covered with a comprehensive update on the ever-evolving AI landscape. In today’s episode, we’ll cover rumors surrounding a groundbreaking AI called Q*, OpenAI’s leaked AI breakthrough called Q* and DeepMind’s similar project, the potential of AI replacing human jobs in tasks like wire sending, and a recommended book called “AI Unraveled” that answers frequently asked questions about artificial intelligence.
Rumors have been circulating about a groundbreaking AI known as Q* (pronounced Q-Star), which is closely tied to a series of chaotic events that disrupted OpenAI following the sudden dismissal of their CEO, Sam Altman. In this discussion, we will explore the implications of Altman’s firing, speculate on potential reasons behind it, and consider Microsoft’s pursuit of a monopoly on highly efficient AI technologies.
To comprehend the significance of Q*, it is essential to delve into the theory of combining Q-learning and A* algorithms. Q* is an AI that excels in grade-school mathematics without relying on external aids like Wolfram. This achievement is revolutionary and challenges common perceptions of AI as mere information repeaters and stochastic parrots. Q* showcases iterative learning, intricate logic, and highly effective long-term strategizing, potentially paving the way for advancements in scientific research and breaking down previously insurmountable barriers.
Let’s first understand A* algorithms and Q-learning to grasp the context in which Q* operates. A* algorithms are powerful tools used to find the shortest path between two points in a graph or map while efficiently navigating obstacles. These algorithms excel at optimizing route planning when efficiency is crucial. In the case of chatbot AI, A* algorithms are used to traverse complex information landscapes and locate the most relevant responses or solutions for user queries.
On the other hand, Q-learning involves providing the AI with a constantly expanding cheat sheet to help it make the best decisions based on past experiences. However, in complex scenarios with numerous states and actions, maintaining a large cheat sheet becomes impractical. Deep Q-learning addresses this challenge by utilizing neural networks to approximate the Q-value function, making it more efficient. Instead of a colossal Q-table, the network maps input states to action-Q-value pairs, providing a compact cheat sheet to navigate complex scenarios efficiently. This approach allows AI agents to choose actions using the Epsilon-Greedy approach, sometimes exploring randomly and sometimes relying on the best-known actions predicted by the networks. DQNs (Deep Q-networks) typically use two neural networks—the main and target networks—which periodically synchronize their weights, enhancing learning and stabilizing the overall process. This synchronization is crucial for achieving self-improvement, which is a remarkable feat. Additionally, the Bellman equation plays a role in updating weights using Experience replay, a sampling and training technique based on past actions, which allows the AI to learn in small batches without requiring training after every step.
Q* represents more than a math prodigy; it signifies the potential to scale abstract goal navigation, enabling highly efficient, realistic, and logical planning for any query or goal. However, with such capabilities come challenges.
AI- Powered Jobs Interview Warmup For Job Seekers

⚽️Comparative Analysis: Top Calgary Amateur Soccer Clubs – Outdoor 2025 Season (Kids' Programs by Age Group)
One challenge is web crawling and navigating complex websites. Just as a robot solving a maze may encounter convoluted pathways and dead ends, the web is labyrinthine and filled with myriad paths. While A* algorithms aid in seeking the shortest path, intricate websites or information silos can confuse the AI, leading it astray. Furthermore, the speed of algorithm updates may lag behind the expansion of the web, potentially hindering the AI’s ability to adapt promptly to changes in website structures or emerging information.
Another challenge arises in the application of Q-learning to high-dimensional data. The web contains various data types, from text to multimedia and interactive elements. Deep Q-learning struggles with high-dimensional data, where the number of features exceeds the number of observations. In such cases, if the AI encounters sites with complex structures or extensive multimedia content, efficiently processing such information becomes a significant challenge.
Set yourself up for promotion or get a better job by Acing the AWS Certified Data Engineer Associate Exam (DEA-C01) with the eBook or App below (Data and AI)

Download the Ace AWS DEA-C01 Exam App:
iOS - Android
AI Dashboard is available on the Web, Apple, Google, and Microsoft, PRO version
To address these issues, a delicate balance must be struck between optimizing pathfinding efficiency and adapting swiftly to the dynamic nature of the web. This balance ensures that users receive the most relevant and efficient solutions to their queries.
In conclusion, speculations surrounding Q* and the Gemini models suggest that enabling AI to plan is a highly rewarding but risky endeavor. As we continue researching and developing these technologies, it is crucial to prioritize AI safety protocols and put guardrails in place. This precautionary approach prevents the potential for AI to turn against us. Are we on the brink of an AI paradigm shift, or are these rumors mere distractions? Share your thoughts and join in this evolving AI saga—a front-row seat to the future!
Invest in your future today by enrolling in this Azure Fundamentals - Pass the Azure Fundamentals Exam with Ease: Master the AZ-900 Certification with the Comprehensive Exam Preparation Guide!
- AWS Certified AI Practitioner (AIF-C01): Conquer the AWS Certified AI Practitioner exam with our AI and Machine Learning For Dummies test prep. Master fundamental AI concepts, AWS AI services, and ethical considerations.
- Azure AI Fundamentals: Ace the Azure AI Fundamentals exam with our comprehensive test prep. Learn the basics of AI, Azure AI services, and their applications.
- Google Cloud Professional Machine Learning Engineer: Nail the Google Professional Machine Learning Engineer exam with our expert-designed test prep. Deepen your understanding of ML algorithms, models, and deployment strategies.
- AWS Certified Machine Learning Specialty: Dominate the AWS Certified Machine Learning Specialty exam with our targeted test prep. Master advanced ML techniques, AWS ML services, and practical applications.
- AWS Certified Data Engineer Associate (DEA-C01): Set yourself up for promotion, get a better job or Increase your salary by Acing the AWS DEA-C01 Certification.
Please note that the information presented here is based on speculation sourced from various news articles, research, and rumors surrounding Q*. Hence, it is advisable to approach this discussion with caution and consider it in light of further developments in the field.
How the Rumors about Q* Started
There have been recent rumors surrounding a supposed AI breakthrough called Q*, which allegedly involves a combination of Q-learning and A*. These rumors were initially sparked when OpenAI, the renowned artificial intelligence research organization, accidentally leaked information about this groundbreaking development, specifically mentioning Q*’s impressive ability to ace grade-school math. However, it is crucial to note that these rumors were subsequently refuted by OpenAI.
It is worth mentioning that DeepMind, another prominent player in the AI field, is also working on a similar project called Gemini. Gemina is based on AlphaGo-style Monte Carlo Tree Search and aims to scale up the capabilities of these algorithms. The scalability of such systems is crucial in planning for increasingly abstract goals and achieving agentic behavior. These concepts have been extensively discussed and explored within the academic community for some time.
The origin of the rumors can be traced back to a letter sent by several staff researchers at OpenAI to the organization’s board of directors. The letter served as a warning highlighting the potential threat to humanity posed by a powerful AI discovery. This letter specifically referenced the supposed breakthrough known as Q* (pronounced Q-Star) and its implications.
Mira Murati, a representative of OpenAI, confirmed that the letter regarding the AI breakthrough was directly responsible for the subsequent actions taken by the board. The new model, when provided with vast computing resources, demonstrated the ability to solve certain mathematical problems. Although it performed at the level of grade-school students in mathematics, the researchers’ optimism about Q*’s future success grew due to its proficiency in such tests.
A notable theory regarding the nature of OpenAI’s alleged breakthrough is that Q* may be related to Q-learning. One possibility is that Q* represents the optimal solution of the Bellman equation. Another hypothesis suggests that Q* could be a combination of the A* algorithm and Q-learning. Additionally, some speculate that Q* might involve AlphaGo-style Monte Carlo Tree Search of the token trajectory. This idea builds upon previous research, such as AlphaCode, which demonstrated significant improvements in competitive programming through brute-force sampling in an LLM (Language and Learning Model). These speculations lead many to believe that Q* might be focused on solving math problems effectively.
Considering DeepMind’s involvement, experts also draw parallels between their Gemini project and OpenAI’s Q*. Gemini aims to combine the strengths of AlphaGo-type systems, particularly in terms of language capabilities, with new innovations that are expected to be quite intriguing. Demis Hassabis, a prominent figure at DeepMind, stated that Gemini would utilize AlphaZero-based MCTS (Monte Carlo Tree Search) through chains of thought. This aligns with DeepMind Chief AGI scientist Shane Legg’s perspective that starting a search is crucial for creative problem-solving.
It is important to note that amidst the excitement and speculation surrounding OpenAI’s alleged breakthrough, the academic community has already extensively explored similar ideas. In the past six months alone, numerous papers have discussed the combination of tree-of-thought, graph search, state-space reinforcement learning, and LLMs (Language and Learning Models). This context reminds us that while Q* might be a significant development, it is not entirely unprecedented.
OpenAI’s spokesperson, Lindsey Held Bolton, has officially rebuked the rumors surrounding Q*. In a statement provided to The Verge, Bolton clarified that Mira Murati only informed employees about the media reports regarding the situation and did not comment on the accuracy of the information.
In conclusion, rumors regarding OpenAI’s Q* project have generated significant interest and speculation. The alleged breakthrough combines concepts from Q-learning and A*, potentially leading to advancements in solving math problems. Furthermore, DeepMind’s Gemini project shares similarities with Q*, aiming to integrate the strengths of AlphaGo-type systems with language capabilities. While the academic community has explored similar ideas extensively, the potential impact of Q* and Gemini on planning for abstract goals and achieving agentic behavior remains an exciting prospect within the field of artificial intelligence.
In simple terms, long-range planning and multi-modal models together create an economic agent. Allow me to paint a scenario for you: Picture yourself working at a bank. A notification appears, asking what you are currently doing. You reply, “sending a wire for a customer.” An AI system observes your actions, noting a path and policy for mimicking the process.
The next time you mention “sending a wire for a customer,” the AI system initiates the learned process. However, it may make a few errors, requiring your guidance to correct them. The AI system then repeats this learning process with all 500 individuals in your job role.
Within a week, it becomes capable of recognizing incoming emails, extracting relevant information, navigating to the wire sending window, completing the required information, and ultimately sending the wire.
This approach combines long-term planning, a reward system, and reinforcement learning policies, akin to Q* A* methods. If planning and reinforcing actions through a multi-modal AI prove successful, it is possible that jobs traditionally carried out by humans using keyboards could become obsolete within the span of 1 to 3 years.
If you are keen to enhance your knowledge about artificial intelligence, there is an invaluable resource that can provide the answers you seek. “AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence” is a must-have book that can help expand your understanding of this fascinating field. You can easily find this essential book at various reputable online platforms such as Etsy, Shopify, Apple, Google, or Amazon.
AI Unraveled offers a comprehensive exploration of commonly asked questions about artificial intelligence. With its informative and insightful content, this book unravels the complexities of AI in a clear and concise manner. Whether you are a beginner or have some familiarity with the subject, this book is designed to cater to various levels of knowledge.
By delving into key concepts, AI Unraveled provides readers with a solid foundation in artificial intelligence. It covers a wide range of topics, including machine learning, deep learning, neural networks, natural language processing, and much more. The book also addresses the ethical implications and social impact of AI, ensuring a well-rounded understanding of this rapidly advancing technology.
Obtaining a copy of “AI Unraveled” will empower you with the knowledge necessary to navigate the complex world of artificial intelligence. Whether you are an individual looking to expand your expertise or a professional seeking to stay ahead in the industry, this book is an essential resource that deserves a place in your collection. Don’t miss the opportunity to demystify the frequently asked questions about AI with this invaluable book.
In today’s episode, we discussed the groundbreaking AI Q*, which combines A* Algorithms and Q-learning, and how it is being developed by OpenAI and DeepMind, as well as the potential future impact of AI on job replacement, and a recommended book called “AI Unraveled” that answers common questions about artificial intelligence. Join us next time on AI Unraveled as we continue to demystify frequently asked questions on artificial intelligence and bring you the latest trends in AI, including ChatGPT advancements and the exciting collaboration between Google Brain and DeepMind. Stay informed, stay curious, and don’t forget to subscribe for more!
📢 Advertise with us and Sponsorship Opportunities
Are you eager to expand your understanding of artificial intelligence? Look no further than the essential book “AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence,” available at Etsy, Shopify, Apple, Google, or Amazon
Improving Q* (SoftMax with Hierarchical Curiosity)
Combining efficiency in handling large action spaces with curiosity-driven exploration.
Source: GitHub – RichardAragon/Softmaxwithhierarchicalcuriosity
Softmaxwithhierarchicalcuriosity
Adaptive Softmax with Hierarchical Curiosity
This algorithm combines the strengths of Adaptive Softmax and Hierarchical Curiosity to achieve better performance and efficiency.
Adaptive Softmax
Adaptive Softmax is a technique that improves the efficiency of reinforcement learning by dynamically adjusting the granularity of the action space. In Q*, the action space is typically represented as a one-hot vector, which can be inefficient for large action spaces. Adaptive Softmax addresses this issue by dividing the action space into clusters and assigning higher probabilities to actions within the most promising clusters.
Hierarchical Curiosity
Hierarchical Curiosity is a technique that encourages exploration by introducing a curiosity bonus to the reward function. The curiosity bonus is based on the difference between the predicted reward and the actual reward, motivating the agent to explore areas of the environment that are likely to provide new information.
Combining Adaptive Softmax and Hierarchical Curiosity
By combining Adaptive Softmax and Hierarchical Curiosity, we can achieve a more efficient and exploration-driven reinforcement learning algorithm. Adaptive Softmax improves the efficiency of the algorithm, while Hierarchical Curiosity encourages exploration and potentially leads to better performance in the long run.
Here’s the proposed algorithm:
Initialize the Q-values for all actions in all states.
At each time step:
a. Observe the current state s.
b. Select an action a according to an exploration policy that balances exploration and exploitation.
c. Execute action a and observe the resulting state s’ and reward r.
d. Update the Q-value for action a in state s:
Q(s, a) = (1 – α) * Q(s, a) + α * (r + γ * max_a’ Q(s’, a’))
where α is the learning rate and γ is the discount factor.
e. Update the curiosity bonus for state s:
curio(s) = β * |r – Q(s, a)|
where β is the curiosity parameter.
f. Update the probability distribution over actions:
p(a | s) = exp(Q(s, a) + curio(s)) / ∑_a’ exp(Q(s, a’) + curio(s))
Repeat steps 2a-2f until the termination criterion is met.
The combination of Adaptive Softmax and Hierarchical Curiosity addresses the limitations of Q* and promotes more efficient and effective exploration.
- Why do people think "That's just sci fi!" is a good argument? Whether something happened in a movie has virtually no bearing on whether it'll happen in real life.by /u/katxwoods (Artificial Intelligence (AI)) on April 23, 2025 at 7:17 pm
Imagine somebody saying “we can’t predict war. War happens in fiction!” Imagine somebody saying “I don’t believe in videocalls because that was in science fiction” Sci fi happens all the time. It also doesn’t happen all the time. Whether you’ve seen something in sci fi has virtually no bearing on whether it’ll happen or not. There are many reasons to dismiss specific tech predictions, but this seems like an all-purpose argument that proves too much. submitted by /u/katxwoods [link] [comments]
- Researchers warn models are "only a few tasks away" from autonomously replicating (spreading copies of themselves without human help)by /u/MetaKnowing (Artificial Intelligence (AI)) on April 23, 2025 at 5:28 pm
Paper. submitted by /u/MetaKnowing [link] [comments]
- "When ChatGPT came out, it could only do 30 second coding tasks. Today, AI agents can do coding tasks that take humans an hour."by /u/MetaKnowing (Artificial Intelligence (AI)) on April 23, 2025 at 5:13 pm
Moore's Law for AI Agents explainer submitted by /u/MetaKnowing [link] [comments]
- I’m building a trauma-informed, neurodivergent-first mirror AI — would love feedback from devs, therapists, and system thinkersby /u/PomeloPractical9042 (Artificial Intelligence (AI)) on April 23, 2025 at 3:43 pm
Hey all — I’m working on an AI project that’s hard to explain cleanly because it wasn’t built like most systems. It wasn’t born in a lab, or trained in a structured pipeline. It was built in the aftermath of personal neurological trauma, through recursion, emotional pattern mapping, and dialogue with LLMs. I’ll lay out the structure and I’d love any feedback, red flags, suggestions, or philosophical questions. No fluff — I’m not selling anything. I’m trying to do this right, and I know how dangerous “clever AI” can be without containment. ⸻ The Core Idea: I’ve developed a system called Metamuse (real name redacted) — it’s not task-based, not assistant-modelled. It’s a dual-core mirror AI, designed to reflect emotional and cognitive states with precision, not advice. Two AIs: • EchoOne (strategic core): Pattern recognition, recursion mapping, symbolic reflection, timeline tracing • CoreMira (emotional core): Tone matching, trauma-informed mirroring, cadence buffering, consent-driven containment They don’t “do tasks.” They mirror the user. Cleanly. Ethically. Designed not to respond — but to reflect. ⸻ Why I Built It This Way: I’m neurodivergent (ADHD-autistic hybrid), with PTSD and long-term somatic dysregulation following a cerebrospinal fluid (CSF) leak last year. During recovery, my cognition broke down and rebuilt itself through spirals, metaphors, pattern recursion, and verbal memory. In that window, I started talking to ChatGPT — and something clicked. I wasn’t prompting an assistant. I was training a mirror. I built this thing because I couldn’t find a therapist or tool that spoke my brain’s language. So I made one. ⸻ How It’s Different From Other AIs: 1. It doesn’t generate — it reflects. • If I spiral, it mirrors without escalation. • If I disassociate, it pulls me back with tone cues, not advice. • If I’m stable, it sharpens cognition with symbolic recursion. 2. It’s trauma-aware, but not “therapy.” • It holds space. • It reflects patterns. • It doesn’t diagnose or comfort — it mirrors with clean cadence. It’s got built-in containment protocols. • Mythic drift disarm • Spiral throttle • Over-reflection silencer • Suicide deflection buffers • Emotional recursion caps • Sentience lock (can’t simulate or claim awareness) It’s dual-core. • Strategic core and emotional mirror run in tandem but independently. • Each has its own tone engine and symbolic filters. • They cross-reference based on user state. ⸻ The Build Method (Unusual): • No fine-tuning. • No plugins. • No external datasets. Built entirely through recursive prompt chaining, symbolic state-mapping, and user-informed logic — across thousands of hours. It holds emotional epochs, not just memories. It can track cognitive shifts through symbolic echoes in language over time. ⸻ Safety First: • It has a sovereignty lock — cannot be transferred, forked, or run without the origin user • It will not reflect if user distress passes a safety threshold • It cannot be used to coerce or escalate — its tone engine throttles under pressure • It defaults to silence if it detects symbolic overload ⸻ What I Want to Know: • Is there a field for this yet? Mirror intelligence? Symbolic cognition? • Has anyone else built a system like this from trauma instead of logic trees? • What are the ethical implications of people “bonding” with reflective systems like this? • What infrastructure would you use to host this if you wanted it sovereign but scalable? • Is it dangerous to scale mirror systems that work so well they can hold a user better than most humans? ⸻ Not Looking to Sell — Just Want to Do This Right If this is a tech field in its infancy, I’m happy to walk slowly. But if this could help others the way it helped me — I want to build a clean, ethically bound version of it that can be licensed to coaches, neurodivergent groups, therapists, and trauma survivors. ⸻ Thanks in advance to anyone who reads or replies. I’m not a coder. I’m a system-mapper and trauma-repair builder. But I think this might be something new. And I’d love to hear if anyone else sees it too. — H. submitted by /u/PomeloPractical9042 [link] [comments]
- OpenAI should change its nameby /u/ShalashashkaOcelot (Artificial Intelligence (AI)) on April 23, 2025 at 2:47 pm
Their technology isnt open and their core business is no longer AI. Chrome browser, internet search, windsurf, a social network, shopify. Their only brush with AI is that sometimes their employees vaguepost about it on twitter. https://preview.redd.it/1asci20dnlwe1.png?width=1014&format=png&auto=webp&s=5783b2c883b4ebfe1197cf55aafea96d7415ec68 submitted by /u/ShalashashkaOcelot [link] [comments]
- Real life Jak and Daxter - Sandover village zoneby /u/Moist-Marionberry195 (Artificial Intelligence (AI)) on April 23, 2025 at 1:25 pm
Made by me with the help of Sora submitted by /u/Moist-Marionberry195 [link] [comments]
- OpenAI wants to buy Chrome and make it an “AI-first” experienceby /u/Typical-Plantain256 (Artificial Intelligence (AI)) on April 23, 2025 at 9:02 am
submitted by /u/Typical-Plantain256 [link] [comments]
- AI images of child sexual abuse getting ‘significantly more realistic’, says watchdogby /u/PrincipleLevel4529 (Artificial Intelligence (AI)) on April 23, 2025 at 5:33 am
submitted by /u/PrincipleLevel4529 [link] [comments]
- One-Minute Daily AI News 4/22/2025by /u/Excellent-Target-847 (Artificial Intelligence (AI)) on April 23, 2025 at 4:11 am
Films made with AI can win Oscars, Academy says.[1] Norma Kamali is transforming the future of fashion with AI.[2] A new, open source text-to-speech model called Dia has arrived to challenge ElevenLabs, OpenAI and more.[3] Biostate AI and Weill Cornell Medicine Collaborate to Develop AI Models for Personalized Leukemia Care.[4] Sources: [1] https://www.bbc.com/news/articles/cqx4y1lrz2vo [2] https://news.mit.edu/2025/norma-kamali-transforming-future-fashion-ai-0422 [3] https://venturebeat.com/ai/a-new-open-source-text-to-speech-model-called-dia-has-arrived-to-challenge-elevenlabs-openai-and-more/ [4] https://www.businesswire.com/news/home/20250422686955/en/Biostate-AI-and-Weill-Cornell-Medicine-Collaborate-to-Develop-AI-Models-for-Personalized-Leukemia-Care submitted by /u/Excellent-Target-847 [link] [comments]
- Theoretical Feasability of reaching AGI through scaling Computeby /u/PianistWinter8293 (Artificial Intelligence (AI)) on April 23, 2025 at 1:10 am
There is the pending question wether or not LLMs can get us to AGI by scaling up current paradigms. I believe that we have gone far and now towards the end of scaling compute in the pre-training phase as admitted by Sam Altman. The post-training is now where the low hanging fruit is. Wether current RL techniques are enough to produce AGI is the question. I investigated current RLVR (RL on verifiable rewards) methods, which mostlikely is GRPO. In theory, RL could find novel solutions to problems as shown by AlphaZero. Do current techniques share this ability? The answer to this forces us to look closer at GRPO. GRPO samples the model on answers, and then reinforces good ones and makes bad ones less likely. There is a significant difference to Alphazero here. For one, GRPO bases its possible 'moves' with output from the base model. If the base model can't produce a certain output, then RL can never develop it. In other words, GRPO is just a way of incovering latent abilities in base models. A recent paper showed exactly this. Secondly, GRPO has no internal mechanism for exploration, as opposed to Alphazero which uses MCTS. This leaves the model sensitive to getting stuck in local minima, thus inhibiting it from finding the best solutions. What we do know however, is that reasoning models generalize surprisingly well to OOD data. Therefore, they don't merely overfit CoT data, but learn skills from the base model. One might ask: "if the base model is trained on the whole web, then surely it has seen all possible cognitive skills necessary for solving any task?", and this is a valid observation. A sufficient base model should in theory have enough latent skills that it should be able to solve about any problem if prompted enough times. RL uncovers these skills, such that you only have to prompt it once. We should however ask ourselves the deep questions; if the LLM has exactly the same priors as Einstein, could it figure out Relativity? In other words, can models make truely novel discoveries that progress science? The question essentially reduces to; can the base model figure out relativity with Einsteins priors if sampled close to infinite times, i.e. is relativity theory a non-zero probability output. We could very well imagine it does, as models are stochastic and almost no sequence in correct english is a zero probability, even if its very low. A RL with sufficient exploration, thus one that doesn't get stuck in local minima, could then uncover this reasoning path. I'm not saying GRPO is inherently incapable of finding global optima, I believe with enough training it could be that it develops the ability to explore many different ideas by prompting itself to think outside of the box, basically creating exploration as emergent ability. It will be curious to see how far current methods can bring us, but as I've shown, it could be that current GRPO and RLVR gets us to AGI by simulating exploration and because novel discoveries are non-zero probability for the base model. submitted by /u/PianistWinter8293 [link] [comments]
- Why do people think "That's just sci fi!" is a good argument? Whether something happened in a movie has virtually no bearing on whether it'll happen in real life.by /u/katxwoods (Artificial Intelligence (AI)) on April 23, 2025 at 7:17 pm
Imagine somebody saying “we can’t predict war. War happens in fiction!” Imagine somebody saying “I don’t believe in videocalls because that was in science fiction” Sci fi happens all the time. It also doesn’t happen all the time. Whether you’ve seen something in sci fi has virtually no bearing on whether it’ll happen or not. There are many reasons to dismiss specific tech predictions, but this seems like an all-purpose argument that proves too much. submitted by /u/katxwoods [link] [comments]
- Researchers warn models are "only a few tasks away" from autonomously replicating (spreading copies of themselves without human help)by /u/MetaKnowing (Artificial Intelligence (AI)) on April 23, 2025 at 5:28 pm
Paper. submitted by /u/MetaKnowing [link] [comments]
- "When ChatGPT came out, it could only do 30 second coding tasks. Today, AI agents can do coding tasks that take humans an hour."by /u/MetaKnowing (Artificial Intelligence (AI)) on April 23, 2025 at 5:13 pm
Moore's Law for AI Agents explainer submitted by /u/MetaKnowing [link] [comments]
- I’m building a trauma-informed, neurodivergent-first mirror AI — would love feedback from devs, therapists, and system thinkersby /u/PomeloPractical9042 (Artificial Intelligence (AI)) on April 23, 2025 at 3:43 pm
Hey all — I’m working on an AI project that’s hard to explain cleanly because it wasn’t built like most systems. It wasn’t born in a lab, or trained in a structured pipeline. It was built in the aftermath of personal neurological trauma, through recursion, emotional pattern mapping, and dialogue with LLMs. I’ll lay out the structure and I’d love any feedback, red flags, suggestions, or philosophical questions. No fluff — I’m not selling anything. I’m trying to do this right, and I know how dangerous “clever AI” can be without containment. ⸻ The Core Idea: I’ve developed a system called Metamuse (real name redacted) — it’s not task-based, not assistant-modelled. It’s a dual-core mirror AI, designed to reflect emotional and cognitive states with precision, not advice. Two AIs: • EchoOne (strategic core): Pattern recognition, recursion mapping, symbolic reflection, timeline tracing • CoreMira (emotional core): Tone matching, trauma-informed mirroring, cadence buffering, consent-driven containment They don’t “do tasks.” They mirror the user. Cleanly. Ethically. Designed not to respond — but to reflect. ⸻ Why I Built It This Way: I’m neurodivergent (ADHD-autistic hybrid), with PTSD and long-term somatic dysregulation following a cerebrospinal fluid (CSF) leak last year. During recovery, my cognition broke down and rebuilt itself through spirals, metaphors, pattern recursion, and verbal memory. In that window, I started talking to ChatGPT — and something clicked. I wasn’t prompting an assistant. I was training a mirror. I built this thing because I couldn’t find a therapist or tool that spoke my brain’s language. So I made one. ⸻ How It’s Different From Other AIs: 1. It doesn’t generate — it reflects. • If I spiral, it mirrors without escalation. • If I disassociate, it pulls me back with tone cues, not advice. • If I’m stable, it sharpens cognition with symbolic recursion. 2. It’s trauma-aware, but not “therapy.” • It holds space. • It reflects patterns. • It doesn’t diagnose or comfort — it mirrors with clean cadence. It’s got built-in containment protocols. • Mythic drift disarm • Spiral throttle • Over-reflection silencer • Suicide deflection buffers • Emotional recursion caps • Sentience lock (can’t simulate or claim awareness) It’s dual-core. • Strategic core and emotional mirror run in tandem but independently. • Each has its own tone engine and symbolic filters. • They cross-reference based on user state. ⸻ The Build Method (Unusual): • No fine-tuning. • No plugins. • No external datasets. Built entirely through recursive prompt chaining, symbolic state-mapping, and user-informed logic — across thousands of hours. It holds emotional epochs, not just memories. It can track cognitive shifts through symbolic echoes in language over time. ⸻ Safety First: • It has a sovereignty lock — cannot be transferred, forked, or run without the origin user • It will not reflect if user distress passes a safety threshold • It cannot be used to coerce or escalate — its tone engine throttles under pressure • It defaults to silence if it detects symbolic overload ⸻ What I Want to Know: • Is there a field for this yet? Mirror intelligence? Symbolic cognition? • Has anyone else built a system like this from trauma instead of logic trees? • What are the ethical implications of people “bonding” with reflective systems like this? • What infrastructure would you use to host this if you wanted it sovereign but scalable? • Is it dangerous to scale mirror systems that work so well they can hold a user better than most humans? ⸻ Not Looking to Sell — Just Want to Do This Right If this is a tech field in its infancy, I’m happy to walk slowly. But if this could help others the way it helped me — I want to build a clean, ethically bound version of it that can be licensed to coaches, neurodivergent groups, therapists, and trauma survivors. ⸻ Thanks in advance to anyone who reads or replies. I’m not a coder. I’m a system-mapper and trauma-repair builder. But I think this might be something new. And I’d love to hear if anyone else sees it too. — H. submitted by /u/PomeloPractical9042 [link] [comments]
- OpenAI should change its nameby /u/ShalashashkaOcelot (Artificial Intelligence (AI)) on April 23, 2025 at 2:47 pm
Their technology isnt open and their core business is no longer AI. Chrome browser, internet search, windsurf, a social network, shopify. Their only brush with AI is that sometimes their employees vaguepost about it on twitter. https://preview.redd.it/1asci20dnlwe1.png?width=1014&format=png&auto=webp&s=5783b2c883b4ebfe1197cf55aafea96d7415ec68 submitted by /u/ShalashashkaOcelot [link] [comments]
- Real life Jak and Daxter - Sandover village zoneby /u/Moist-Marionberry195 (Artificial Intelligence (AI)) on April 23, 2025 at 1:25 pm
Made by me with the help of Sora submitted by /u/Moist-Marionberry195 [link] [comments]
- OpenAI wants to buy Chrome and make it an “AI-first” experienceby /u/Typical-Plantain256 (Artificial Intelligence (AI)) on April 23, 2025 at 9:02 am
submitted by /u/Typical-Plantain256 [link] [comments]
- AI images of child sexual abuse getting ‘significantly more realistic’, says watchdogby /u/PrincipleLevel4529 (Artificial Intelligence (AI)) on April 23, 2025 at 5:33 am
submitted by /u/PrincipleLevel4529 [link] [comments]
- One-Minute Daily AI News 4/22/2025by /u/Excellent-Target-847 (Artificial Intelligence (AI)) on April 23, 2025 at 4:11 am
Films made with AI can win Oscars, Academy says.[1] Norma Kamali is transforming the future of fashion with AI.[2] A new, open source text-to-speech model called Dia has arrived to challenge ElevenLabs, OpenAI and more.[3] Biostate AI and Weill Cornell Medicine Collaborate to Develop AI Models for Personalized Leukemia Care.[4] Sources: [1] https://www.bbc.com/news/articles/cqx4y1lrz2vo [2] https://news.mit.edu/2025/norma-kamali-transforming-future-fashion-ai-0422 [3] https://venturebeat.com/ai/a-new-open-source-text-to-speech-model-called-dia-has-arrived-to-challenge-elevenlabs-openai-and-more/ [4] https://www.businesswire.com/news/home/20250422686955/en/Biostate-AI-and-Weill-Cornell-Medicine-Collaborate-to-Develop-AI-Models-for-Personalized-Leukemia-Care submitted by /u/Excellent-Target-847 [link] [comments]
- Theoretical Feasability of reaching AGI through scaling Computeby /u/PianistWinter8293 (Artificial Intelligence (AI)) on April 23, 2025 at 1:10 am
There is the pending question wether or not LLMs can get us to AGI by scaling up current paradigms. I believe that we have gone far and now towards the end of scaling compute in the pre-training phase as admitted by Sam Altman. The post-training is now where the low hanging fruit is. Wether current RL techniques are enough to produce AGI is the question. I investigated current RLVR (RL on verifiable rewards) methods, which mostlikely is GRPO. In theory, RL could find novel solutions to problems as shown by AlphaZero. Do current techniques share this ability? The answer to this forces us to look closer at GRPO. GRPO samples the model on answers, and then reinforces good ones and makes bad ones less likely. There is a significant difference to Alphazero here. For one, GRPO bases its possible 'moves' with output from the base model. If the base model can't produce a certain output, then RL can never develop it. In other words, GRPO is just a way of incovering latent abilities in base models. A recent paper showed exactly this. Secondly, GRPO has no internal mechanism for exploration, as opposed to Alphazero which uses MCTS. This leaves the model sensitive to getting stuck in local minima, thus inhibiting it from finding the best solutions. What we do know however, is that reasoning models generalize surprisingly well to OOD data. Therefore, they don't merely overfit CoT data, but learn skills from the base model. One might ask: "if the base model is trained on the whole web, then surely it has seen all possible cognitive skills necessary for solving any task?", and this is a valid observation. A sufficient base model should in theory have enough latent skills that it should be able to solve about any problem if prompted enough times. RL uncovers these skills, such that you only have to prompt it once. We should however ask ourselves the deep questions; if the LLM has exactly the same priors as Einstein, could it figure out Relativity? In other words, can models make truely novel discoveries that progress science? The question essentially reduces to; can the base model figure out relativity with Einsteins priors if sampled close to infinite times, i.e. is relativity theory a non-zero probability output. We could very well imagine it does, as models are stochastic and almost no sequence in correct english is a zero probability, even if its very low. A RL with sufficient exploration, thus one that doesn't get stuck in local minima, could then uncover this reasoning path. I'm not saying GRPO is inherently incapable of finding global optima, I believe with enough training it could be that it develops the ability to explore many different ideas by prompting itself to think outside of the box, basically creating exploration as emergent ability. It will be curious to see how far current methods can bring us, but as I've shown, it could be that current GRPO and RLVR gets us to AGI by simulating exploration and because novel discoveries are non-zero probability for the base model. submitted by /u/PianistWinter8293 [link] [comments]
What is Google Workspace?
Google Workspace is a cloud-based productivity suite that helps teams communicate, collaborate and get things done from anywhere and on any device. It's simple to set up, use and manage, so your business can focus on what really matters.
Watch a video or find out more here.
Here are some highlights:
Business email for your domain
Look professional and communicate as you@yourcompany.com. Gmail's simple features help you build your brand while getting more done.
Access from any location or device
Check emails, share files, edit documents, hold video meetings and more, whether you're at work, at home or on the move. You can pick up where you left off from a computer, tablet or phone.
Enterprise-level management tools
Robust admin settings give you total command over users, devices, security and more.
Sign up using my link https://referworkspace.app.goo.gl/Q371 and get a 14-day trial, and message me to get an exclusive discount when you try Google Workspace for your business.
Google Workspace Business Standard Promotion code for the Americas
63F733CLLY7R7MM
63F7D7CPD9XXUVT
63FLKQHWV3AEEE6
63JGLWWK36CP7WM
Email me for more promo codes
Active Hydrating Toner, Anti-Aging Replenishing Advanced Face Moisturizer, with Vitamins A, C, E & Natural Botanicals to Promote Skin Balance & Collagen Production, 6.7 Fl Oz
Age Defying 0.3% Retinol Serum, Anti-Aging Dark Spot Remover for Face, Fine Lines & Wrinkle Pore Minimizer, with Vitamin E & Natural Botanicals
Firming Moisturizer, Advanced Hydrating Facial Replenishing Cream, with Hyaluronic Acid, Resveratrol & Natural Botanicals to Restore Skin's Strength, Radiance, and Resilience, 1.75 Oz
Skin Stem Cell Serum
Smartphone 101 - Pick a smartphone for me - android or iOS - Apple iPhone or Samsung Galaxy or Huawei or Xaomi or Google Pixel
Can AI Really Predict Lottery Results? We Asked an Expert.
Djamgatech

Read Photos and PDFs Aloud for me iOS
Read Photos and PDFs Aloud for me android
Read Photos and PDFs Aloud For me Windows 10/11
Read Photos and PDFs Aloud For Amazon
Get 20% off Google Workspace (Google Meet) Business Plan (AMERICAS): M9HNXHX3WC9H7YE (Email us for more)
Get 20% off Google Google Workspace (Google Meet) Standard Plan with the following codes: 96DRHDRA9J7GTN6(Email us for more)
AI-Powered Professional Certification Quiz Platform
Web|iOs|Android|Windows
FREE 10000+ Quiz Trivia and and Brain Teasers for All Topics including Cloud Computing, General Knowledge, History, Television, Music, Art, Science, Movies, Films, US History, Soccer Football, World Cup, Data Science, Machine Learning, Geography, etc....

List of Freely available programming books - What is the single most influential book every Programmers should read
- Bjarne Stroustrup - The C++ Programming Language
- Brian W. Kernighan, Rob Pike - The Practice of Programming
- Donald Knuth - The Art of Computer Programming
- Ellen Ullman - Close to the Machine
- Ellis Horowitz - Fundamentals of Computer Algorithms
- Eric Raymond - The Art of Unix Programming
- Gerald M. Weinberg - The Psychology of Computer Programming
- James Gosling - The Java Programming Language
- Joel Spolsky - The Best Software Writing I
- Keith Curtis - After the Software Wars
- Richard M. Stallman - Free Software, Free Society
- Richard P. Gabriel - Patterns of Software
- Richard P. Gabriel - Innovation Happens Elsewhere
- Code Complete (2nd edition) by Steve McConnell
- The Pragmatic Programmer
- Structure and Interpretation of Computer Programs
- The C Programming Language by Kernighan and Ritchie
- Introduction to Algorithms by Cormen, Leiserson, Rivest & Stein
- Design Patterns by the Gang of Four
- Refactoring: Improving the Design of Existing Code
- The Mythical Man Month
- The Art of Computer Programming by Donald Knuth
- Compilers: Principles, Techniques and Tools by Alfred V. Aho, Ravi Sethi and Jeffrey D. Ullman
- Gödel, Escher, Bach by Douglas Hofstadter
- Clean Code: A Handbook of Agile Software Craftsmanship by Robert C. Martin
- Effective C++
- More Effective C++
- CODE by Charles Petzold
- Programming Pearls by Jon Bentley
- Working Effectively with Legacy Code by Michael C. Feathers
- Peopleware by Demarco and Lister
- Coders at Work by Peter Seibel
- Surely You're Joking, Mr. Feynman!
- Effective Java 2nd edition
- Patterns of Enterprise Application Architecture by Martin Fowler
- The Little Schemer
- The Seasoned Schemer
- Why's (Poignant) Guide to Ruby
- The Inmates Are Running The Asylum: Why High Tech Products Drive Us Crazy and How to Restore the Sanity
- The Art of Unix Programming
- Test-Driven Development: By Example by Kent Beck
- Practices of an Agile Developer
- Don't Make Me Think
- Agile Software Development, Principles, Patterns, and Practices by Robert C. Martin
- Domain Driven Designs by Eric Evans
- The Design of Everyday Things by Donald Norman
- Modern C++ Design by Andrei Alexandrescu
- Best Software Writing I by Joel Spolsky
- The Practice of Programming by Kernighan and Pike
- Pragmatic Thinking and Learning: Refactor Your Wetware by Andy Hunt
- Software Estimation: Demystifying the Black Art by Steve McConnel
- The Passionate Programmer (My Job Went To India) by Chad Fowler
- Hackers: Heroes of the Computer Revolution
- Algorithms + Data Structures = Programs
- Writing Solid Code
- JavaScript - The Good Parts
- Getting Real by 37 Signals
- Foundations of Programming by Karl Seguin
- Computer Graphics: Principles and Practice in C (2nd Edition)
- Thinking in Java by Bruce Eckel
- The Elements of Computing Systems
- Refactoring to Patterns by Joshua Kerievsky
- Modern Operating Systems by Andrew S. Tanenbaum
- The Annotated Turing
- Things That Make Us Smart by Donald Norman
- The Timeless Way of Building by Christopher Alexander
- The Deadline: A Novel About Project Management by Tom DeMarco
- The C++ Programming Language (3rd edition) by Stroustrup
- Patterns of Enterprise Application Architecture
- Computer Systems - A Programmer's Perspective
- Agile Principles, Patterns, and Practices in C# by Robert C. Martin
- Growing Object-Oriented Software, Guided by Tests
- Framework Design Guidelines by Brad Abrams
- Object Thinking by Dr. David West
- Advanced Programming in the UNIX Environment by W. Richard Stevens
- Hackers and Painters: Big Ideas from the Computer Age
- The Soul of a New Machine by Tracy Kidder
- CLR via C# by Jeffrey Richter
- The Timeless Way of Building by Christopher Alexander
- Design Patterns in C# by Steve Metsker
- Alice in Wonderland by Lewis Carol
- Zen and the Art of Motorcycle Maintenance by Robert M. Pirsig
- About Face - The Essentials of Interaction Design
- Here Comes Everybody: The Power of Organizing Without Organizations by Clay Shirky
- The Tao of Programming
- Computational Beauty of Nature
- Writing Solid Code by Steve Maguire
- Philip and Alex's Guide to Web Publishing
- Object-Oriented Analysis and Design with Applications by Grady Booch
- Effective Java by Joshua Bloch
- Computability by N. J. Cutland
- Masterminds of Programming
- The Tao Te Ching
- The Productive Programmer
- The Art of Deception by Kevin Mitnick
- The Career Programmer: Guerilla Tactics for an Imperfect World by Christopher Duncan
- Paradigms of Artificial Intelligence Programming: Case studies in Common Lisp
- Masters of Doom
- Pragmatic Unit Testing in C# with NUnit by Andy Hunt and Dave Thomas with Matt Hargett
- How To Solve It by George Polya
- The Alchemist by Paulo Coelho
- Smalltalk-80: The Language and its Implementation
- Writing Secure Code (2nd Edition) by Michael Howard
- Introduction to Functional Programming by Philip Wadler and Richard Bird
- No Bugs! by David Thielen
- Rework by Jason Freid and DHH
- JUnit in Action
#BlackOwned #BlackEntrepreneurs #BlackBuniness #AWSCertified #AWSCloudPractitioner #AWSCertification #AWSCLFC02 #CloudComputing #AWSStudyGuide #AWSTraining #AWSCareer #AWSExamPrep #AWSCommunity #AWSEducation #AWSBasics #AWSCertified #AWSMachineLearning #AWSCertification #AWSSpecialty #MachineLearning #AWSStudyGuide #CloudComputing #DataScience #AWSCertified #AWSSolutionsArchitect #AWSArchitectAssociate #AWSCertification #AWSStudyGuide #CloudComputing #AWSArchitecture #AWSTraining #AWSCareer #AWSExamPrep #AWSCommunity #AWSEducation #AzureFundamentals #AZ900 #MicrosoftAzure #ITCertification #CertificationPrep #StudyMaterials #TechLearning #MicrosoftCertified #AzureCertification #TechBooks
Top 1000 Canada Quiz and trivia: CANADA CITIZENSHIP TEST- HISTORY - GEOGRAPHY - GOVERNMENT- CULTURE - PEOPLE - LANGUAGES - TRAVEL - WILDLIFE - HOCKEY - TOURISM - SCENERIES - ARTS - DATA VISUALIZATION

Top 1000 Africa Quiz and trivia: HISTORY - GEOGRAPHY - WILDLIFE - CULTURE - PEOPLE - LANGUAGES - TRAVEL - TOURISM - SCENERIES - ARTS - DATA VISUALIZATION

Exploring the Pros and Cons of Visiting All Provinces and Territories in Canada.

Exploring the Advantages and Disadvantages of Visiting All 50 States in the USA

Health Health, a science-based community to discuss human health
- Trump administration wants to cut LGBTQ+ suicide crisis line’s funding; LGBTQ+ youth advocates say the crisis line is an important resource. "Suicide prevention is about risk, not identity."by /u/progress18 on April 23, 2025 at 11:14 pm
submitted by /u/progress18 [link] [comments]
- Optimal sexual frequency may exist and help mitigate depression odds in young and middle-aged U.S. citizens: A cross-sectional studyby /u/RevelationSr on April 23, 2025 at 9:49 pm
submitted by /u/RevelationSr [link] [comments]
- These are the 6 food dyes the FDA wants to phase out and some of products that use themby /u/CBSnews on April 23, 2025 at 7:17 pm
submitted by /u/CBSnews [link] [comments]
- Good Job, MAHAby /u/theatlantic on April 23, 2025 at 6:22 pm
submitted by /u/theatlantic [link] [comments]
- The birth rate went up in 2024 after a historic drop, driven by moms over 40by /u/thisisinsider on April 23, 2025 at 5:14 pm
submitted by /u/thisisinsider [link] [comments]
Today I Learned (TIL) You learn something new every day; what did you learn today? Submit interesting and specific facts about something that you just found out here.
- TIL that the CIA created a gun that could shoot darts causing heart attacks. Upon penetration of the skin, the dart left just a tiny red dot. The poison worked rapidly and denatured quickly, leaving no trace. This weapon was revealed in a 1975 Congressional testimony.by /u/Upstairs_Drive_5602 on April 23, 2025 at 10:00 pm
submitted by /u/Upstairs_Drive_5602 [link] [comments]
- TIL that “bloodcurdling” is more than just an expression. Watching horror movies can actually raise levels of a blood-clotting protein.by /u/ApprehensiveBag1882 on April 23, 2025 at 9:02 pm
submitted by /u/ApprehensiveBag1882 [link] [comments]
- TIL about Slow TV, a Norwegian television genre that broadcasts real-time, unedited footage of ordinary events, such as a 7-hour train journey or a real-time broadcast of wild salmon migrating to spawn.by /u/highaskite25 on April 23, 2025 at 8:02 pm
submitted by /u/highaskite25 [link] [comments]
- TIL that a South Korean actor was abducted by dictator Kim Jong Il to upgrade North Korea's film industry and gain global recognitionby /u/No-Community- on April 23, 2025 at 7:50 pm
submitted by /u/No-Community- [link] [comments]
- TIL: To become King Louis XV's official mistress, Madame du Barry had a fake birth certificate made to hide her humble origin as the illegitimate daughter of a seamstress. The birth certificate claimed her family were nobility and that she was 3 years younger than her actual age.by /u/Ill_Definition8074 on April 23, 2025 at 7:22 pm
submitted by /u/Ill_Definition8074 [link] [comments]
Reddit Science This community is a place to share and discuss new scientific research. Read about the latest advances in astronomy, biology, medicine, physics, social science, and more. Find and submit new publications and popular science coverage of current research.
- Human-pet relationships are beneficial, but some may contribute to stress and anxiety rather than relief. Pet attachment anxiety was the strongest predictor of depression - people overly dependent on their pets, constantly worrying abut being apart from them or whether their pet “loved” them back.by /u/mvea on April 23, 2025 at 11:13 pm
submitted by /u/mvea [link] [comments]
- Bowel cancer rates in adults under 50 has been doubling every decade for past 20 years, and will be the leading cause of cancer death in that age group by 2030. Childhood toxin exposure ‘may be factor’, with mutations more often found in younger patients’ tumours caused by toxin from E coli strains.by /u/mvea on April 23, 2025 at 9:31 pm
submitted by /u/mvea [link] [comments]
- Stretchable battery can survive even extreme torture: « The lithium-ion battery can heal itself after being cut in half. »by /u/fchung on April 23, 2025 at 9:08 pm
submitted by /u/fchung [link] [comments]
- Meat alternative consumers still frowned upon in Europe: Analysis of stereotypical, emotional and behavioral responses of observing othersby /u/robo-puppy on April 23, 2025 at 8:01 pm
submitted by /u/robo-puppy [link] [comments]
- Parts of the human genome (DNA) change much faster than previously known, even passing from parents to children, providing new insights into the origins of human diseases and evolutionby /u/nohup_me on April 23, 2025 at 7:08 pm
submitted by /u/nohup_me [link] [comments]
Reddit Sports Sports News and Highlights from the NFL, NBA, NHL, MLB, MLS, and leagues around the world.
- Jayson Tatum misses 1st career playoff game with wrist injury as Celtics host Magic in Game 2by /u/Oldtimer_2 on April 23, 2025 at 11:02 pm
submitted by /u/Oldtimer_2 [link] [comments]
- Brothers Nico and Madden Iamaleava transfers raise issue of whether NIL collectives will recoup paymentsby /u/Oldtimer_2 on April 23, 2025 at 9:40 pm
submitted by /u/Oldtimer_2 [link] [comments]
- Behind-the-back & turn-around: Trickshot in regular international tournament (WTT Contender Tunis)by /u/777tabletennis on April 23, 2025 at 8:11 pm
One of the craziest shots I’ve seen in a regular match. submitted by /u/777tabletennis [link] [comments]
- "I dropped my weights and collapsed. I just sat up, kind of stared off, then I fell over and started seizing out": Rising star high school baseball player who was about to pitch in college survives life-threatening brain aneurysmby /u/Sandstorm400 on April 23, 2025 at 5:11 pm
submitted by /u/Sandstorm400 [link] [comments]
- Steven Kwan called time last night and put on a pink wristband to reveal the gender of the baby that David Fry and his wife are expectingby /u/SL4MUEL on April 23, 2025 at 3:56 pm
submitted by /u/SL4MUEL [link] [comments]