[AINews] Bespoke-Stratos + Sky-T1: The Vicuna+Alpaca moment for reasoning
This is AI News! an MVP of a service that goes thru all AI discords/Twitters/reddits and summarizes what people are talking about, so that you can keep up without the fatigue. Signing up here opts you in to the real thing when we launch it 🔜
Reasoning Distillation is all you need.
AI News for 1/21/2025-1/22/2025. We checked 7 subreddits, 433 Twitters and 34 Discords (225 channels, and 4297 messages) for you. Estimated reading time saved (at 200wpm): 496 minutes. You can now tag @smol_ai for AINews discussions!
In the ChatGPT heyday of 2022-23, Alpaca and Vicuna were born out of LMsys and Stanford as ultra cheap ($300) finetunes of LLaMA 1 that distilled from ChatGPT/Bard samples to achieve 90% of the quality of ChatGPT/GPT3.5.
In the last 48 hours, it seems the Berkeley/USC folks have done it again, this time with the reasoning models.
It's hard to believe this sequence of events happened just in the last 2 weeks:
- Berkeley's Sky Computing lab released Sky-T1-32B-Preview, a finetune of Qwen 2.5 32B (our coverage here) with 17k rows of training data from QwQ-32B (our coverage here) + rewriting traces with gpt-4o-mini + rejection sampling, all done for $450. Because QwQ outperforms o1-preview, distilling from QwQ brings Qwen up to match o1-preview's benchmarks:
- DeepSeek releases R1 (2 days ago) with benchmarks far above o1-preview. The R1 paper also revealed the surprise that you can distill from R1 to turn a 1.5B Qwen model to match 4o and 3.5 Sonnet (?!).
- Bespoke Labs (today) uses the Sky-T1 recipe to distill R1 on Qwen again to greatly outperform (not just match) o1-preview, again with 17k rows of reasoning traces.
While Bespoke's distillation does not quite match DeepSeek's distillation in performance, they used 17k samples vs DeepSeek's 800k. It is pretty evident that they could keep going here if they wished.
The bigger shocking thing is that "SFT is all you need" - no major architecture changes are required for reasoning to happen, just feed in more (validated, rephrased) reasoning traces, backtracking and pivoting and all, and it seems like it will generalize well. In all likelihood, this explains the relative efficiency of o1-mini and o3-mini vs their full size counterparts.
Table of Contents
- AI Twitter Recap
- AI Reddit Recap
- AI Discord Recap
- PART 1: High level Discord summaries
- Unsloth AI (Daniel Han) Discord
- Codeium (Windsurf) Discord
- LM Studio Discord
- Nous Research AI Discord
- Interconnects (Nathan Lambert) Discord
- aider (Paul Gauthier) Discord
- Stackblitz (Bolt.new) Discord
- Yannick Kilcher Discord
- OpenRouter (Alex Atallah) Discord
- Perplexity AI Discord
- MCP (Glama) Discord
- Latent Space Discord
- GPU MODE Discord
- Eleuther Discord
- OpenAI Discord
- Notebook LM Discord Discord
- Stability.ai (Stable Diffusion) Discord
- LlamaIndex Discord
- Modular (Mojo 🔥) Discord
- LAION Discord
- Cohere Discord
- LLM Agents (Berkeley MOOC) Discord
- Nomic.ai (GPT4All) Discord
- DSPy Discord
- Torchtune Discord
- OpenInterpreter Discord
- Gorilla LLM (Berkeley Function Calling) Discord
- tinygrad (George Hotz) Discord
- Axolotl AI Discord
- Mozilla AI Discord
- PART 2: Detailed by-Channel summaries and links
- Unsloth AI (Daniel Han) ▷ #general (266 messages🔥🔥):
- Unsloth AI (Daniel Han) ▷ #off-topic (6 messages):
- Unsloth AI (Daniel Han) ▷ #help (146 messages🔥🔥):
- Unsloth AI (Daniel Han) ▷ #research (191 messages🔥🔥):
- Codeium (Windsurf) ▷ #discussion (48 messages🔥):
- Codeium (Windsurf) ▷ #windsurf (426 messages🔥🔥🔥):
- LM Studio ▷ #announcements (1 messages):
- LM Studio ▷ #general (209 messages🔥🔥):
- LM Studio ▷ #hardware-discussion (180 messages🔥🔥):
- Nous Research AI ▷ #general (295 messages🔥🔥):
- Nous Research AI ▷ #ask-about-llms (6 messages):
- Nous Research AI ▷ #research-papers (5 messages):
- Nous Research AI ▷ #interesting-links (12 messages🔥):
- Nous Research AI ▷ #research-papers (5 messages):
- Nous Research AI ▷ #reasoning-tasks (1 messages):
- Interconnects (Nathan Lambert) ▷ #news (87 messages🔥🔥):
- Interconnects (Nathan Lambert) ▷ #ml-drama (16 messages🔥):
- Interconnects (Nathan Lambert) ▷ #random (88 messages🔥🔥):
- Interconnects (Nathan Lambert) ▷ #memes (4 messages):
- Interconnects (Nathan Lambert) ▷ #rl (13 messages🔥):
- Interconnects (Nathan Lambert) ▷ #reads (12 messages🔥):
- Interconnects (Nathan Lambert) ▷ #posts (3 messages):
- Interconnects (Nathan Lambert) ▷ #policy (4 messages):
- aider (Paul Gauthier) ▷ #general (122 messages🔥🔥):
- aider (Paul Gauthier) ▷ #questions-and-tips (91 messages🔥🔥):
- aider (Paul Gauthier) ▷ #links (1 messages):
- Stackblitz (Bolt.new) ▷ #announcements (1 messages):
- Stackblitz (Bolt.new) ▷ #prompting (16 messages🔥):
- Stackblitz (Bolt.new) ▷ #discussions (192 messages🔥🔥):
- Yannick Kilcher ▷ #general (161 messages🔥🔥):
- Yannick Kilcher ▷ #paper-discussion (32 messages🔥):
- Yannick Kilcher ▷ #agents (1 messages):
- Yannick Kilcher ▷ #ml-news (3 messages):
- OpenRouter (Alex Atallah) ▷ #announcements (1 messages):
- OpenRouter (Alex Atallah) ▷ #general (187 messages🔥🔥):
- Perplexity AI ▷ #announcements (1 messages):
- Perplexity AI ▷ #general (128 messages🔥🔥):
- Perplexity AI ▷ #sharing (8 messages🔥):
- Perplexity AI ▷ #pplx-api (16 messages🔥):
- MCP (Glama) ▷ #general (114 messages🔥🔥):
- MCP (Glama) ▷ #showcase (7 messages):
- Latent Space ▷ #ai-general-chat (96 messages🔥🔥):
- Latent Space ▷ #ai-announcements (1 messages):
- GPU MODE ▷ #general (5 messages):
- GPU MODE ▷ #triton (18 messages🔥):
- GPU MODE ▷ #cuda (16 messages🔥):
- GPU MODE ▷ #torch (6 messages):
- GPU MODE ▷ #beginner (13 messages🔥):
- GPU MODE ▷ #pmpp-book (4 messages):
- GPU MODE ▷ #jax (1 messages):
- GPU MODE ▷ #torchao (7 messages):
- GPU MODE ▷ #off-topic (9 messages🔥):
- GPU MODE ▷ #self-promotion (4 messages):
- GPU MODE ▷ #arc-agi-2 (10 messages🔥):
- Eleuther ▷ #general (26 messages🔥):
- Eleuther ▷ #research (16 messages🔥):
- Eleuther ▷ #lm-thunderdome (29 messages🔥):
- Eleuther ▷ #gpt-neox-dev (2 messages):
- OpenAI ▷ #ai-discussions (44 messages🔥):
- OpenAI ▷ #gpt-4-discussions (2 messages):
- OpenAI ▷ #prompt-engineering (6 messages):
- OpenAI ▷ #api-discussions (6 messages):
- Notebook LM Discord ▷ #use-cases (10 messages🔥):
- Notebook LM Discord ▷ #general (43 messages🔥):
- Stability.ai (Stable Diffusion) ▷ #general-chat (53 messages🔥):
- LlamaIndex ▷ #blog (3 messages):
- LlamaIndex ▷ #general (46 messages🔥):
- Modular (Mojo 🔥) ▷ #general (6 messages):
- Modular (Mojo 🔥) ▷ #mojo (29 messages🔥):
- LAION ▷ #general (34 messages🔥):
- Cohere ▷ #discussions (25 messages🔥):
- Cohere ▷ #questions (1 messages):
- Cohere ▷ #api-discussions (1 messages):
- Cohere ▷ #cmd-r-bot (6 messages):
- LLM Agents (Berkeley MOOC) ▷ #mooc-questions (20 messages🔥):
- Nomic.ai (GPT4All) ▷ #general (17 messages🔥):
- DSPy ▷ #general (9 messages🔥):
- DSPy ▷ #examples (1 messages):
- Torchtune ▷ #dev (8 messages🔥):
- Torchtune ▷ #papers (2 messages):
- OpenInterpreter ▷ #general (5 messages):
- Gorilla LLM (Berkeley Function Calling) ▷ #discussion (3 messages):
- tinygrad (George Hotz) ▷ #learn-tinygrad (2 messages):
- Axolotl AI ▷ #general (1 messages):
- Mozilla AI ▷ #announcements (1 messages):
AI Twitter Recap
all recaps done by Claude 3.5 Sonnet, best of 4 runs.
AI Model Developments and Evaluations
- DeepSeek-R1 Innovations and Performance: @teortaxesTex, @cwolferesearch, and @madiator discussed DeepSeek-R1's training via pure reinforcement learning (RL), emphasizing the importance of supervised finetuning (SFT) for accelerating RL convergence. DeepSeek-R1 demonstrates robust reasoning capabilities and multimodal functionalities, with Bespoke-Stratos-32B introduced as a distilled version achieving significant performance with 47x fewer examples.
- Gemini and Other LLM Advancements: @chakraAI and @philschmid highlighted Google's Gemini 2.0 Flash Thinking model, noting its 1 million token context window, code execution support, and state-of-the-art performance across math, science, and multimodal reasoning benchmarks.
- AI Model Comparisons and Critiques: @abacaj and @teortaxesTex provided critical insights into models like o1 and R1-Zero, discussing issues such as model repeatability, behavioral self-awareness, and the limitations of RLHF in achieving robust reasoning.
AI Applications and Tools
- Windsurf and AI-Powered Slide Decks: @omarsar0 showcased Windsurf, an AI agent capable of analyzing code, replicating functionalities, and automating the creation of slide decks by integrating PDFs and images seamlessly. Users can extend features through simple prompts, highlighting the flexibility of web-based AI applications.
- Local AI Deployments and Extensions: @ggerganov introduced the llama.cpp server, which offers unique context reuse techniques for enhancing LLM completions based on codebase contents, optimized for low-end hardware. Additionally, the VS Code extension leveraging llama.cpp provides local LLM-assisted code and text completions without the need for external RAG systems.
- AI Integration with Development Tools: @lah2139 and @JayMcMillan highlighted integrations like LlamaIndex with DeepSeek-R1, enabling AI-assisted development and agent workflows. These tools allow developers to build and evaluate multi-agent systems, fostering efficient AI application development.
AI Research and Papers
- IntellAgent Multi-Agent Framework: @omarsar0 introduced IntellAgent, an open-source multi-agent framework designed to evaluate complex conversational AI systems. The framework facilitates the generation of synthetic benchmarks and interactive user-agent simulations, capturing the intricate dynamics of agent capabilities and policy constraints.
- Behavioral Self-Awareness in LLMs: @omarsar0 discussed a new paper demonstrating that LLMs can exhibit behavioral self-awareness by recognizing and commenting on their own insecure code outputs without explicit training, indicating a potential for more reliable policy enforcement within models.
- ModernBERT and Embedding Models: @philschmid presented ModernBERT, an embedding and ranking model that correctly associates contextual information better than its predecessors. The comparison revealed that relying solely on benchmarks may not fully capture a model's effectiveness, emphasizing the need for customized evaluation strategies.
AI Infrastructure and Compute
- OpenAI's Stargate Project: @sama and @gdb announced the Stargate Project, a $500 billion AI infrastructure initiative aimed at building AI data centers in the U.S., positioning it as a response to global AI competition and a strategy to enhance national AI capabilities.
- NVIDIA's AI Models and Compute Solutions: @reach_vb detailed NVIDIAAI's Eagle 2, a suite of vision-language models (VLMs) that outperform competitors like GPT-4o on specific benchmarks, underscoring the importance of efficient compute architectures in developing high-performance AI models.
- Compute Resource Management: @swyx and @cto_junior discussed strategies for managing inference-time compute, balancing costs with adversarial robustness, and the implications of compute resource allocation on AI model performance.
AI Community, Education, and Events
- AI Workshops and Courses: @deeplearningai and @AndrewYNg promoted hands-on workshops and free courses focused on building AI agents capable of computer use, covering topics like multimodal prompting, XML structuring, and prompt caching to enhance AI assistant functionalities.
- AI Film Festival Growth: @c_valenzuelab celebrated the expansion of their Film Festival, noting a 10x increase in submissions and the relocation to prominent venues like Alice Tully Hall, reflecting the growing intersection of AI media and creative industries.
- AI Community Contributions: @LangChainAI and @Hacubu showcased community-driven projects like AgentWorkflow and LangSmith Evals, which simplify the process of creating multi-agent systems and testing LLM applications, thereby enhancing community collaboration and developer productivity.
Memes/Humor
- AI Model Humor: @giffmana and @saranormous shared humorous takes on AI model limitations and user interactions, including jokes about chatbot behaviors and AI-driven creativity mishaps.
- Satirical Comments on AI Developments: @nearcyan and @giffmana posted satirical remarks on AI project naming conventions and misunderstandings in AI capabilities, adding a light-hearted perspective to the rapid advancements in the field.
AI Policy and Ethics
- AI Safety and Governance: @togelius expressed concerns over the AI safety agenda, advocating for a balanced approach that prioritizes freedom of compute while addressing existential risks, highlighting the tension between AI innovation and ethical considerations.
- AI Community Critiques: @pthoughtcrime___ and @simran_s_arora critiqued policy-driven AI initiatives, emphasizing the potential for control over AI development and the importance of maintaining open-source principles to foster ethical AI progress.
- Regulatory Discussions: @agihippo and @labloke11 engaged in conversations about the impact of AI regulations on innovation and research, debating the balance between regulatory oversight and technological advancement.
AI Reddit Recap
/r/LocalLlama Recap
Theme 1. Mistral 10V: Exploring New Capabilities with 12K Tokens
- China goes full robotic. Insane developments. At the moment, it’s a heated race between USA and China. (Score: 768, Comments: 237): The post highlights the intense competition between the USA and China in the field of AI, specifically mentioning China's advancements in robotics. The release of Mistral 10V is implied to have significant implications for AI technology, though no specific details are provided in the text.
- Many commenters express skepticism about the authenticity of the video, questioning if it is AI-generated or preprogrammed, with some noting the absence of the video on official channels and others discussing the possibility of it being a fake.
- The discussion reflects a belief that China is significantly ahead in robotics, with some commenters suggesting that the USA has fallen behind due to a focus on non-manufacturing industries, and questioning the notion of a "race" between the two countries.
- There is a mix of humor and concern regarding the future implications of advanced robotics, with comments mentioning potential military applications and the replacement of jobs, such as police dogs, by robots once they achieve further capabilities.
- Ooh... Awkward (Score: 560, Comments: 295): OpenAI launched Mistral 10V with a capacity of 12,000 tokens, marking a significant development in AI capabilities. The context of the post suggests a potential awkward situation, possibly related to the launch or features of the model.
- The vocal fry of the speaker was a prominent topic, with many commenters criticizing it as distracting or unprofessional. Sam Altman's demeanor during the presentation was perceived as lacking confidence, with some speculating that his nervousness stemmed from the context of the situation rather than the content itself.
- Discussions touched on the potential economic implications of AI, with skepticism about claims that AI would create 100,000 jobs. Commenters expressed doubts about job creation, suggesting that AI might instead reduce the workforce, and drew parallels to Theranos as a cautionary tale.
- There were political undertones, with references to Donald Trump and the notion of AI being used to claim achievements like curing cancer. Some commenters suggested that Sam Altman was navigating a complex political landscape, trying to maintain a favorable position amid Trump's influence.
- Sam Altman’s expression during the entire AI Infra Deal Announcement (Score: 469, Comments: 131): The post lacks specific content or discussion points about Sam Altman or the AI Infra Deal Announcement. Without additional context or details, no technical summary can be provided.
- Discussions compare Russia's oligarchic system to the US, noting concerns about increasing alignment of wealthy individuals with Trump, potentially to gain influence. This reflects worries about a shift towards oligarchic tendencies, akin to Russia's political structure under Putin, where oligarchs face severe consequences if they fall out of favor.
- There is commentary on Sam Altman's demeanor during public appearances, with some attributing his expressions to anxiety or discomfort. The reactions suggest skepticism about his enthusiasm for certain partnerships, possibly alluding to a satirical take on him creating a Skynet-like scenario against his will.
- Conversations include sarcastic remarks about Elon Musk being sidelined in AI advancements compared to Altman, with references to Musk's involvement in other initiatives like meme coins. The humor underlines a perceived rivalry or competition in the tech space between influential figures.
Theme 2. O1-Pro: Revolutionary Use in Legislation Analysis
- I used O1-pro to Analyze the Constitutionality of all of Trump's Executive Orders. (Score: 135, Comments: 33): The author conducted a detailed analysis of Trump's Executive Orders using O1-Pro and sourced the texts from whitehouse.gov for objectivity. The document includes a Table of Contents and source text links, with summaries provided by GT4o.
- Analysis Process: The author manually prepared the document, using Google Doc's bookmark and link system for navigation. They used O1-Pro for analysis, inserting the full text of executive orders and employing prompt templates for summaries and titles, ensuring each analysis was conducted in a fresh chat to avoid bias.
- Impact of Executive Orders: Discussion highlighted potential short and long-term effects, such as immediate restructuring within federal agencies, changes in immigration policies, and shifts in energy and environmental focus. Long-term impacts could include a smaller, more centralized federal workforce and shifts in international relations due to withdrawal from treaties.
- Fact-Checking and Economic Concerns: Commenters suggested sharing actual ChatGPT links for fact-checking and speculated on economic effects, like tariffs and their impact on prices in Canada and the US. There was skepticism about whether proposed tariffs would be enacted without an official order.
- [D]: A 3blue1brown Video that Explains Attention Mechanism in Detail (Score: 285, Comments: 12): 3blue1brown's video on the attention mechanism provides a detailed explanation of concepts such as token embedding and the role of the embedding space in encoding multiple meanings for a word. It discusses how a well-trained attention block adjusts embeddings based on context and conceptualizes Ks as potentially answering Qs. Video link and subtitles are provided for further exploration.
- 3blue1brown's video is praised for its clear and visual explanation of the attention mechanism, effectively introducing the problem and gradually building up to the solution, unlike other tutorials that often skip foundational explanations.
- Users highlight the importance of masking during model training to predict the next token, referencing Karpathy's tutorial on building GPT from scratch for further understanding of these concepts.
- 3blue1brown's talk, based on the video series, is also recommended for its intuitive explanations, with a link provided for those interested in exploring more: YouTube link.
- Trump announces up to $500 billion in AI infrastructure investment (Score: 141, Comments: 22): OpenAI, SoftBank, and Oracle are launching a Texas-based joint venture named Stargate, with an initial commitment of $100 billion and plans to invest up to $500 billion over the next four years. This venture aims to set new standards in AI infrastructure.
- Multiple comments highlight that the Stargate project was initially announced last year, with some users expressing skepticism about its recent announcement being politically motivated, particularly regarding Trump taking credit for it.
- There is a discussion about the scale of the project, with one commenter noting it as potentially the largest infrastructure project in history, while another user humorously compares it to Foxconn 2.0.
- Some users express gratitude for the straightforward announcement style, avoiding sensationalist headlines like "BREAKING," which they find increasingly meaningless.
Theme 3. Gemini 1.5: Leading AI with Performance Edge
- Elon Says Softbank Doesn't Have the Funding.. (Score: 417, Comments: 227): Elon Musk expresses doubt about SoftBank's financial capabilities, contradicting claims of substantial funding for AI infrastructure by stating they have "well under $10B secured." An image from OpenAI announces "The Stargate Project," which intends to invest $500 billion in AI infrastructure in the U.S. over four years, starting with an immediate $100 billion deployment.
- There is skepticism about SoftBank's financial claims, with some suggesting they might have announced their plans without securing the necessary funding, hoping investments would follow. Concerns are raised about the legality and ethics of business leaders, like Elon Musk, influencing or commenting on other businesses, especially given his controversial track record and connections to government initiatives.
- Discussions highlight Trump's announced $500 billion government subsidy for AI, drawing parallels to past infrastructure funding that was misused. Critics view it as a potential wealth transfer to tech elites, questioning the involvement and actual financial capabilities of companies like SoftBank and Oracle.
- Many comments express disdain for Elon Musk, questioning his motives and credibility, with accusations of personal biases and trolling. There are references to past unfulfilled promises, such as the XAi Grok model, and criticism for his perceived alignment with controversial figures and ideologies.
- OpenAI announcement on the Stargate project (Score: 186, Comments: 90): OpenAI's Stargate Project plans to invest $500 billion over four years in AI infrastructure in the U.S., starting with an immediate $100 billion deployment. Initial equity funders include SoftBank, OpenAI, Oracle, and MGX, with key technology partners like Arm, Microsoft, NVIDIA, Oracle, and OpenAI. The project aims to support American jobs, national security, and advance AGI for humanity's benefit.
- Comments discuss skepticism about SoftBank's involvement, with questions on why they wouldn't invest in Japan and clarifications that SoftBank is linked to the Middle East sovereign fund. Concerns about the funding sources were raised, noting potential reliance on subsidies and tax credits.
- The discussion highlights confusion over the project’s goals, with some suggesting the $100 billion will be spent on data centers, AI R&D labs, and energy infrastructure, referencing Microsoft's past venture with the Three Mile Island nuclear plant. There is skepticism about the project's employment claims, comparing it to other tech initiatives like Tesla and Alexa.
- The term "Stargate" is humorously compared to "Skynet" from the Terminator series, with some comments noting that a Skynet program already exists as a military satellite system. There is a mention of the project contributing to the "re-industrialization of the United States" and being part of the fourth industrial revolution.
Other AI Subreddit Recap
/r/Singularity, /r/Oobabooga, /r/MachineLearning, /r/OpenAI, /r/ClaudeAI, /r/StableDiffusion, /r/ChatGPT
Theme 1. Stargate AI Project: $500 Billion Investment's Impact
- Trump announces a $500 billion AI infrastructure investment in the US (Score: 582, Comments: 355): Trump announced a $500 billion investment in AI infrastructure in the US, signaling a significant commitment to advancing AI capabilities and infrastructure development. The announcement highlights the growing importance of AI in national economic and technological strategies.
- Stargate Project: The $500 billion investment is directed towards a new private company called Stargate, co-owned by Sam Altman, Masayoshi Son, and Larry Ellison, rather than OpenAI. This has raised concerns about intellectual property and existing partnerships, particularly with Microsoft.
- Funding and Ownership: There is a debate about the source of the funding, with some suggesting it's private investment from companies like SoftBank rather than US government money. Trump's announcement is seen by some as a political move, with claims that the announcement allows him to take credit for private sector initiatives.
- Geopolitical and Economic Implications: The announcement is viewed as a strategic move in the global AI race, especially in response to China's advancements like DeepSeek R1. The discussion also touches on the potential economic impact, including job creation claims and the broader implications for US technological leadership.
- I don’t believe the $500 Billion OpenAI investment (Score: 419, Comments: 142): The post expresses skepticism about the $500 billion OpenAI investment, arguing that the figure is overly optimistic and lacks transparency regarding funding sources and project specifics. The author criticizes the use of vague legal language and suggests the announcement is politically motivated, particularly in timing it after Trump's presidential win, to create headlines without firm commitments, hinting that the actual investment will be smaller and slower than advertised.
- Commenters highlight skepticism regarding the $500 billion investment, with comparisons to past projects like the Foxconn and Star Wars programs, suggesting the announcement is more about stock manipulation and market hype than actual funding. UncannyRobotPodcast and others express concerns about the lack of follow-through and potential for insider trading benefits.
- There is debate over the role of government versus private companies in funding, with tertain and ThreeKiloZero clarifying that the funding is from four partner companies, not federal money, while 05032-MendicantBias notes the government's role in easing regulations for infrastructure. SoftBank is mentioned as a significant player with substantial assets potentially involved.
- Discussions around the potential impact of the investment include concerns about overhyping AI and the existential implications of AI development. NebulousNitrate argues that large investments are justified to prevent adversaries from gaining superintelligence, while Super_Sierra suggests the hype is beneficial for innovation, despite skepticism about the actual realization of the $500 billion goal.
- Just a comparison of US $500B Stargate AI project to other tech projects (Score: 112, Comments: 103): The $500 billion Stargate AI project is compared to historical tech projects, highlighting its scale at approximately 1.7% of US GDP in 2024. In contrast, the Manhattan Project cost about $30 billion (~1.5% of GDP in the 1940s), the Apollo Program around $170–$180 billion (~0.5% of GDP in the 1960s), and the Space Shuttle Program approximately $275–$300 billion (~0.2% of GDP in the 1980s). The Interstate Highway System cost $500–$550 billion over several decades (~0.2%–0.3% of GDP annually).
- Discussion centers around the private funding of the Stargate AI project, with SoftBank, OpenAI, Oracle, and MGX as key investors. There is skepticism about the project's intentions, with comments suggesting it may replace a significant portion of the workforce (10-30%) while contrasting the lack of funding for social welfare programs like healthcare and education in the US.
- The project's scale and impact are debated, with comparisons made to historical projects like the Manhattan Project and Apollo Program regarding their GDP percentages. Some argue that while the project is funded privately, its scale is akin to a public initiative, raising questions about its societal implications and the role of the US government.
- Concerns are voiced about the US's role in AI development, with some commenters expressing distrust in the government's motives and potential exploitation by wealthy interests. There is a sentiment that the US is focused on maintaining global dominance, similar to a new "space race" with China, and that the project could eventually be defense-oriented.
Theme 2. DeepSeek R1: Redefining AI Benchmarks
- R1 is mind blowing (Score: 578, Comments: 139): R1 demonstrated superior problem-solving capabilities in a nuanced graph theory problem, succeeding on the first attempt where 4o failed twice before eventually providing a correct answer. The author is impressed by R1's ability to justify its solution and articulate a nuanced understanding, suggesting that even smaller models running on personal devices like a MacBook may surpass human intelligence in specific areas.
- Users discussed the R1 model's performance compared to o1 and other models, with some emphasizing R1's superior value due to its lower cost despite similar performance levels. Discussions highlighted R1's capabilities in problem-solving and reasoning, with some users noting its impressive performance even in distilled versions.
- R1's problem-solving abilities were praised, with specific examples like successfully solving a graph theory problem on the first attempt and outperforming other models like 4o. However, some users noted limitations, such as a lack of context awareness and issues with prompt optimization.
- The discussion included technical details about model deployment and usage, such as the need for specific temperature settings and questions about self-hosting capabilities. Some users expressed challenges with using R1 in professional settings due to data privacy concerns.
- The Deep Seek R1 glaze is unreal but it’s true. (Score: 63, Comments: 46): The author struggled with a programming issue in a RAG machine for two days, trying various major LLMs without success, including OpenAI's O1 Pro. However, the Deep Seek R1 resolved the problem on its first attempt, leading the author to consider it their new preferred tool for coding, potentially replacing OpenAI Pro.
- There is skepticism about OpenAI's LLMs knowing their own architecture, as users like KriosXVII and gliptic point out that such details are unlikely to be included in their training data. Dan-Boy-Dan criticizes the author's claims as marketing tactics and challenges them to post the problem for others to test with different models.
- a_beautiful_rhind and LostMyOtherAcct69 discuss the difference in AI models' personality and architecture, suggesting that Mixture of Experts (MoE) could be the future of AI due to its efficiency and specialization compared to dense models. ReasonablePossum_ argues that US companies prioritize profit over developing such models, while Caffeine_Monster criticizes the excessive positive bias in AI models as counterproductive.
- Multiple users, including Dan-Boy-Dan and emteedub, request the author to post the specific problem that only the Deep Seek R1 solved, expressing doubt about the claims made and indicating a desire to test it across other models.
- Deepseek-R1 is brittle (Score: 61, Comments: 23): The post discusses the brittleness of Deepseek-R1, highlighting its limitations and strengths. It includes an image link here to support the analysis.
- Prompt Optimization: Users found that Deepseek-R1 performs well in specific scenarios, particularly when using the prompt structure from the R1 paper with a temperature of 0.6 and top p of 0.95, which involves tagging reasoning and answers explicitly in the prompt. This method is also referenced in instructions for o1 models, indicating a common approach across reasoning models.
- Model Brittleness: Deepseek-R1 struggles with tasks requiring creativity or subjective answers, often producing incompatible or excessive outputs, as seen in a test where it suggested an impractical tech stack for an app. However, with practice and precise prompting, users noted an improvement in R1's performance, supporting the post's assertion of its brittleness.
- Comparison with Other Models: The discussion highlights that Deepseek-R1 excels in tasks with a single correct answer, such as coding, but falters compared to other models like Deepseek v3 when dealing with more complex or creative tasks. This suggests that while R1 can be effective, its application needs careful consideration and adjustment based on the task requirements.
Theme 3. Model-Agnostic Reasoning: R1 Techniques
- YOU CAN EXTRACT REASONING FROM R1 AND PASS IT ONTO ANY MODEL (Score: 368, Comments: 101): @skirano on Twitter suggests that you can extract reasoning from deepseek-reasoner and apply it to any model, enhancing its performance, as demonstrated with GPT-3.5 turbo.
- Workflow and Reasoning Techniques: Discussion highlights the use of Chain-of-Thought (CoT) prompting and stepped thinking to enhance model reasoning, with @SomeOddCodeGuy suggesting a two-step workflow using a workflow app to achieve interesting results, as demonstrated in a QwQ simulation. Nixellion adds that prompting models to simulate expert discussions can yield improved outcomes, emphasizing the potential of new CoT techniques.
- Critique and Skepticism: Ok-Parsnip-4826 criticizes the notion of extracting reasoning, arguing that it's merely asking one model to summarize another's thoughts without real benefit, while gus_the_polar_bear counters that LLMs might respond differently to prompts from other LLMs, suggesting potential unexplored interactions. nuclearbananana questions the efficiency of using a secondary model due to potential latency and cost implications.
- Technical Implementation and Tools: SomeOddCodeGuy discusses the technicalities of using Wilmer to facilitate connections between Open WebUI and Ollama, highlighting the potential for creating a containerized setup to enhance workflow management. Additionally, xadiant mentions the possibility of injecting reasoning processes into local models via completions API for enhanced performance.
- The distilled R1 models likely work best in workflows, so now's a great time to learn those if you haven't already! (Score: 49, Comments: 14): The DeepSeek-R1 model, as described in the paper "DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning", performs optimally with zero-shot prompting rather than few-shot prompting. The author emphasizes the importance of using workflows to enhance the performance of reasoning models like R1 and its distilled versions, suggesting a structured approach involving summarization and problem-solving to maximize efficiency and output quality.
- Users discuss challenges with DeepSeek-R1's output format consistency, particularly in generating structured formats like JSON, with one user mentioning the use of langgraph to address these issues. Another user seeks more tips or tricks to improve model performance.
- A commenter notes DeepSeek's own acknowledgment of R1's limitations in function calling and complex tasks compared to DeepSeek-V3, citing the research paper's mention of plans to improve using long Chain-of-Thought (CoT) techniques.
- Some users express interest in having the AI revise prompts to achieve better output, indicating a demand for improved prompt engineering to enhance model responses.
Theme 4. Deepseek R1 GRPO Code: Open-Sourcing Breakthrough
- Deepseek R1 GRPO code open sourced 🤯 (Score: 260, Comments: 6): Deepseek R1 has open-sourced its GRPO code, accompanied by a detailed flowchart illustrating the model's components. The diagram highlights sections like Prompts, Completions, Rewards and Advantages, Policy and Reference Policy, and DKL, all of which contribute to a central "objective" through a series of structured calculations involving mean and standard deviation.
- The Deepseek R1 code shared is not the actual R1 code but rather a preference optimization method used in the R1 training process, highlighting the novelty of the RL environment over the reward model in PO training.
- According to the paper, the RL environment employs a straightforward algorithm that evaluates the reasoning start and end token pairs, comparing the model's output with the ground truth from a math dataset.
- A link to the relevant code is available on GitHub.
- DeepSeek-R1-Distill-Qwen-1.5B running 100% locally in-browser on WebGPU. Reportedly outperforms GPT-4o and Claude-3.5-Sonnet on math benchmarks (28.9% on AIME and 83.9% on MATH). (Score: 170, Comments: 38): DeepSeek-R1-Distill-Qwen-1.5B reportedly runs entirely locally in-browser using WebGPU and surpasses GPT-4o and Claude-3.5-Sonnet in math benchmarks, achieving 28.9% on AIME and 83.9% on MATH.
- DeepSeek-R1-Distill-Qwen-1.5B has generated excitement for running entirely locally on WebGPU and outperforming GPT-4o in reasoning tasks. The model is accessible via an online demo and its source code is available on GitHub.
- Discussion around ONNX highlighted its role as a general-purpose ML model and weights storage format, with its ability to convert and transfer models across different engines. GGUF was noted as being optimized for quantized transformer model weights, primarily used with llamacpp and its derivatives.
- The conversation included an analogy comparing file formats like ONNX and GGUF to container formats such as tar, zip, and 7z, emphasizing that they are different layouts for data storage that cater to specific hardware/software preferences.
Theme 5. R1-Zero: AI Reinforcement Learning Breakthroughs
- R1-Zero: Pure RL Creates a Mind We Can’t Decode—Is This AGI’s Dark Mirror? (Score: 204, Comments: 105): DeepSeek-R1-Zero, an AI model developed through pure reinforcement learning (RL) without supervised fine-tuning, has achieved a dramatic increase in AIME math scores from 15.6% to 86.7%, yet its reasoning remains uninterpretable, producing garbled outputs. While its sibling, R1, uses some supervised data to maintain readability, R1-Zero raises concerns about AI alignment and the potential democratization of superintelligence due to its lower API costs—50 times cheaper than OpenAI's—despite its unreadable logic.
- The gibberish outputs of R1-Zero might be a form of symbolic reasoning, where tokens are repurposed beyond their linguistic meaning to convey complex interrelationships. This concept parallels human use of slang or jargon, suggesting that the model's reasoning could be misunderstood as gibberish due to shifts in token semantics, similar to generational language differences.
- Discussion highlights the potential of R1-Zero's reinforcement learning to reinvent token semantics, allowing it to outperform models like R1 that rely on supervised fine-tuning. This raises questions about how to measure safety and alignment in models using a new form of symbolic reasoning, as well as how this might contribute to multimodal AI development.
- There is skepticism about R1-Zero's capabilities, with some commenters suggesting that its outputs are mere errors or hallucinations rather than groundbreaking insights. Others mention the need for more concrete examples and reports to substantiate claims of its reasoning abilities, pointing to Karpathy's predictions and referencing concepts like the "Coconut" paper by Meta for further context.
- Gemini Thinking experimental 01-21 is out! (Score: 71, Comments: 17): The Gemini 2.0 Flash Thinking Experimental model, showcased in the Google AI Studio interface, features advanced options such as "Model," "Token count," and "Temperature" settings. The interface, with its dark theme, allows users to input mathematical problems and adjust various tools and settings for experimentation.
- Gemini 1.5/1.0 received criticism for being rushed and underwhelming, but the new AI Studio models are praised for their improvements, suggesting a more thoughtful development process. Users appreciate the experimental models' availability for testing, wishing other companies would do the same.
- Open weight models are noted to be outpacing closed ones, sparking discussions on their advantages. There is a mention of naming inconsistencies within the model versions, which could cause confusion.
- The Flash Thinking Experimental model has been updated from a 32k to a 1 million context window, aligning with other Google models. Users report mixed experiences regarding speed, with some finding it impressive and others not.
AI Discord Recap
A summary of Summaries of Summaries
o1-preview-2024-09-12
Theme 1. AI's Billion-Dollar Stargate Projects: Lofty Goals and Skepticism
- OpenAI Announces $500B Stargate Project, Skeptics Abound: OpenAI unveiled the Stargate Project, aiming to invest $500 billion in AI infrastructure over four years, but critics like Elon Musk doubt the feasibility of funding, calling it "ridiculous".
- Microsoft and Oracle Pledge Support Amid Funding Doubts: Microsoft CEO confirms commitment to the project saying, "I’m good for my $80 Billion", while Oracle joins in amidst questions about SoftBank's ability to contribute significantly.
- Trump's Stargate Sparks Debate on Tech and Government: President Trump announced Project Stargate, stirring discussions on government involvement in AI and concerns about corporate overreach in tech investments.
Theme 2. AI Models Clash: DeepSeek R1 Outperforms the Giants
- DeepSeek R1 Trumps Gemini and O1 in Math Tasks: Community members praised DeepSeek R1 for achieving 92% in math performance, outperforming models like Gemini and O1, showcasing advanced reasoning capabilities.
- Bespoke-Stratos-32B Emerges as a Reasoning Powerhouse: A new model, Bespoke-Stratos-32B, distilled from DeepSeek-R1, outperforms Sky-T1 and o1-preview in reasoning tasks using only 800k training examples costing $800.
- Gemini 2.0 Flash Thinking Soars to #1: Gemini-2.0-Flash-Thinking claimed the top spot in Chatbot Arena with a 73.3% math score and a 1 million token context window, sparking excitement and comparisons with DeepSeek models.
Theme 3. Censorship vs. Uncensored AI Models: Users Seek Freedom
- DeepSeek R1 Faces Performance Drop Amid Censorship Concerns: Users reported an 85% performance drop in DeepSeek R1 overnight, suspecting increased censorship filters and seeking workarounds for sensitive prompts.
- Frustration with Over-Censored AI Models Grows: Community members mocked heavily censored models like Phi-3.5, expressing that excessive restrictions make models impractical for technical tasks and roleplaying.
- Hunt for Uncensored Models Intensifies: Users discussed favorite uncensored models like Dolphin and Hermes on platforms like OpenRouter, emphasizing the demand for more open AI experiences.
Theme 4. New AI Tools and Innovations Empower Developers
- LlamaIndex Launches AgentWorkflow for Multi-Agent Magic: AgentWorkflow was introduced, enabling developers to build multi-agent systems with expanded tool support, hailed as "the next step for more powerful agent coordination."
- Ai2 ScholarQA Revolutionizes Literature Reviews: Ai2 ScholarQA launched a RAG-based solution for multi-paper queries, helping researchers conduct in-depth literature reviews with cross-referencing capabilities.
- OpenAI's Operator Prepares to Take Actions in Your Browser: Reports indicate OpenAI is prepping Operator, a ChatGPT feature to perform browser actions on behalf of users, raising both excitement and privacy discussions.
Theme 5. AI Development Challenges: From Quantization to Privacy
- Model Quantization Debates and Innovations: Discussions in the Unsloth AI Discord highlighted dynamic 4-bit quantization methods that retain accuracy while reducing VRAM usage, sparking comparisons with BnB 8-bit approaches.
- Privacy Concerns Over AI Data Handling Policies: Users questioned data handling practices of AI services like Codeium and Windsurf, scrutinizing privacy policies and the use of user data for training.
- AI's Role in Cybersecurity Remains Underexplored: Members highlighted a lack of emphasis on AI cybersecurity solutions, noting that companies like CrowdStrike have used ML for years, and suggesting generative AI could automate threat detection and code-based intrusion analysis.
o1-2024-12-17
Theme 1. AI Infrastructure & Funding Frenzy
- Stargate Summons $500B AI Bonanza: President Trump announced a colossal $500B “Stargate Project” for AI data centers, with initial outlays from SoftBank, Oracle, and MGX. Microsoft blog posts call it the largest AI initiative ever, fueling jobs and American AI leadership.
- Musk, SoftBank, Oracle Stir Controversy: Elon Musk scoffed that SoftBank lacks the money, while skeptics questioned SoftBank’s liquidity and debt. Nonetheless, official statements highlight bold optimism around this mega-scale investment.
- Google Bets Another $1B on Anthropic: Google sunk another billion into Anthropic, reinforcing the fierce competition among AI titans. Speculation abounds on rolling funding strategies and Anthropic’s expanding next-gen offerings.
Theme 2. LLM Showdowns & Math Marvels
- DeepSeek R1 Dominates Math: DeepSeek R1 reportedly hits 92% in math performance and outclasses O1 and Gemini in advanced reasoning tasks. Users praise its geometric insights and thorough multi-stage RL training.
- Gemini 2.0 Flash Shoots #1: Google’s Gemini 2.0 Flash-Thinking rockets to the top of Chatbot Arena, boasting a 73.3% math score and a 1 million token context window. Developers anticipate more refined iterations soon.
- Bespoke-Stratos-32B Surges Past Rivals: Distilled from DeepSeek-R1, this model trounces Sky-T1 and o1-preview while requiring 47x fewer examples. The $800 open-source dataset it used spurs interest in cost-effective community data curation.
Theme 3. Reinforcement Learning & GRPO Talk
- Tiny GRPO Gains Real Traction: Devs are running minimal GRPO code for math tasks and praising the simplified approach for easy experimentation. Early adapters note fast iteration cycles and straightforward debugging.
- Kimi-k1.5 Paper Delves Deep: Researchers highlight curriculum-based RL and a length penalty for better model performance. Community feedback prompts newcomers to blend these ideas into new RL training recipes.
- GRPO Debates Spark KL Divergence Drama: Hugging Face users question how GRPO handles KL in advantage calculations. Code contributors weigh the pros and cons of applying KL directly to the loss rather than the rewards.
Theme 4. HPC & GPU Codegen Adventures
- Blackwell Breaks the Code: NVIDIA’s Blackwell B100/B200 codegen 10.0 and RTX 50 codegen 12.0 updates stir anticipation for sm_100a, sm_101a, and maybe sm_120. Community members await an official whitepaper but get by on partial PR notes.
- Triton Tussles with 3.2: INT8×INT8 dot products crash with the new TMA approach, prompting manual fixes and jit refactors. PyTorch Issue #144103 spotlights backward-compat troubles with AttrsDescriptor removal.
- Accel-Sim Talk Stirs Enthusiasm: HPC explorers eye the Accel-Sim framework for GPU emulation on CPUs. A scheduled talk in late March promises deeper insights into simulated GPU performance and code optimization.
Theme 5. RAG Systems & Tool Innovations
- AgentWorkflow Powers Multi-Agents: LlamaIndex unveils a high-level framework that orchestrates parallel tool usage and agent coordination. Enthusiasts see it as the “next step” for robust multi-agent solutions.
- Sonar Pro Surfs SimpleQA: Perplexity’s new Sonar Pro API outperforms rivals in real-time search-grounded Q&A while promising lower costs. Zoom integrates it for AI Companion 2.0, and devs praise its citation-friendly approach.
- Ai2 ScholarQA Bolsters Literature: This system answers multi-paper queries with cross-referencing superpowers, speeding up academic reviews. Researchers can shift from scanning one PDF at a time to gleaning curated insights across entire corpora.
DeepSeek v3
Theme 1. DeepSeek R1 Model Performance and Integration
- DeepSeek R1 Outperforms Competitors in Math and Reasoning: The DeepSeek R1 Distill Qwen 32B model has been praised for its superior performance in complex math tasks, outperforming models like Llama 405B and DeepSeek V3 671B. Users highlighted its geometric reasoning and multi-stage RL training, with benchmarks showing 92% math performance.
- DeepSeek R1 Integration Challenges: Users reported difficulties integrating DeepSeek R1 into platforms like GPT4All and LM Studio, citing missing public model catalogs and the need for llama.cpp updates. Some also faced issues with API performance drops and censorship filters.
- DeepSeek R1's Chain-of-Thought Reasoning: The model's ability to externalize reasoning steps, such as spelling 'razzberry' versus 'raspberry', was highlighted as a unique feature. Users noted its potential for enhancing reasoning in other models like Claude and O1.
Theme 2. AI Model Quantization and Fine-Tuning
- Unsloth Introduces Dynamic 4-bit Quantization: Unsloth's new dynamic 4-bit quantization method selectively avoids compressing certain parameters, maintaining accuracy while reducing VRAM usage. Users compared it favorably to BnB 8-bit, noting its efficiency in model optimization.
- Fine-Tuning Challenges with Phi-4: Users reported issues with fine-tuning the Phi-4 model, citing poor output quality and looping fixes. Suggestions included adjusting LoRA settings and ensuring high-quality datasets for better results.
- Chinchilla Formula Revisited: Discussions around the Chinchilla formula emphasized the balance between model size and training tokens. Users noted that many models exceed optimal thresholds, leading to inefficiencies in resource utilization.
Theme 3. AI Infrastructure and Large-Scale Investments
- Project Stargate: $500B AI Infrastructure Plan: OpenAI's Stargate Project aims to invest $500 billion over four years to build advanced AI infrastructure in the US. Initial funding includes $100 billion from SoftBank, Oracle, and MGX, with a focus on job creation and national security.
- Google Invests $1B in Anthropic: Google's renewed investment in Anthropic signals confidence in the company's next-gen models, fueling speculation about rolling funding strategies and AI competition.
- DeepSeek R1 Hardware Requirements: Running the full DeepSeek R1 671B model requires significant hardware, with estimates suggesting $18,000 for multiple NVIDIA Digits. Users debated the feasibility of such setups, with some opting for cheaper alternatives like 4xA4000.
Theme 4. AI for Creative and Technical Applications
- Gemini 2.0 Flash Thinking Model Launches: Google's Gemini 2.0 Flash Thinking model, with a 1 million token context window and 64K max output tokens, has been tested against DeepSeek R1. Users praised its potential for large-scale reasoning tasks, though some noted challenges in handling complex prompts.
- NotebookLM for Church Services and Study Workflows: A user leveraged NotebookLM to analyze 16 five-hour YouTube livestreams, generating a 250-page book and a 2000-page bible study. Others integrated it into their study routines, praising its efficiency in reference lookups.
- AI Art Faces Hostile Reactions: Users reported negative responses to AI-generated art, with some being told to kill myself for using these tools. This reflects ongoing societal resistance to AI in creative fields.
Theme 5. AI Safety, Ethics, and Regulation
- Concerns Over AI Job Displacement: A founder expressed moral dilemmas over potential layoffs caused by their AI startup's success, sparking debates about the socio-economic impact of AI advancements. Users compared it to everyday automations that also cut jobs.
- AI Safety Index and Model Alignment: Discussions around the AI Safety Index highlighted the need for robust safety metrics in models like MiniCPM. Users questioned the absence of alignment and safety practices in some models, emphasizing the importance of ethical AI development.
- Regulatory Challenges in AI: A session on AI regulations explored the implications of recent policies, with Dylan Patel from SemiAnalysis discussing the winners and losers in the evolving regulatory landscape. Concerns were raised about government-corporate overlap in AI development.
DeepSeek R1
Theme 1. Model Optimization Wars: Quantization, Fine-Tuning, and Scaling Battles
- Unsloth’s Dynamic 4-bit Quantization Shakes Up VRAM Efficiency: Unsloth’s dynamic 4-bit quantization avoids compressing critical parameters, slashing VRAM use while preserving accuracy. Users compared it to BnB 8-bit, noting dynamic 4-bit’s balance of memory and performance.
- Phi-4 Fine-Tuning Fiasco Sparks Architecture Debates: Fine-tuning Phi-4 led to poor outputs, with users blaming model architecture. Workarounds in other notebooks hinted at the need for specialized hyperparameters.
- Chinchilla Scaling Laws Revisited Amid Training Token Excess: Discussions reignited about Chinchilla’s model-size-to-token ratios, with many models overshooting thresholds. Empirical scaling proofs urged resource optimization for efficiency.
Theme 2. AI Infrastructure Arms Race: $500B Projects and Hardware Hurdles
- Project Stargate’s $500B Ambition Faces Funding Skepticism: OpenAI’s Stargate Project aims to deploy $500B for AI infrastructure, but critics like Gavin Baker questioned SoftBank’s liquidity. Elon Musk dismissed the proposal as unrealistic.
- DeepSeek R1 671B Demands $18K Hardware, Sparks Cost Debates: Running DeepSeek R1 671B reportedly requires 4x NVIDIA Digits ($18K), while users proposed cheaper 4xA4000 setups. Skeptics argued bigger models ≠ better ROI.
- Apple Silicon Pushes 4-bit Quantization Limits for R1 32B: M3 Max MacBook Pro users tested 4-bit quantization for R1 32B, noting quality drops. MLX-optimized variants offered speed gains despite precision trade-offs.
Theme 3. Agentic AI: Hype vs. Reality in Autonomous Systems
- “Agentic AI” Label slapped on Scripted Workflows, Users Revolt: Members mocked marketing claims equating basic scripted tools to agentic AI, citing lack of true autonomy. Calls for transparency in “agent” definitions intensified.
- GRPO and T1 RL Papers Signal Reinforcement Learning Shakeup: GRPO’s minimal implementation and T1 RL’s scaled self-verification paper sparked interest. Concerns arose about KL divergence handling in advantage calculations.
- OpenAI’s “Operator” Teases Browser Automation, But API Absent: Leaks revealed ChatGPT’s Operator for browser actions, but no API support. Users speculated it’s a stopgap for “PhD-level Super Agents.”
Theme 4. Tooling Turbulence: IDE Wars, API Quirks, and RAG Realities
- Windsurf’s Auto-Memory Feature Clashes with Cascade’s Loops: Windsurf’s project memory won praise, but Cascade botched FastAPI file access and looped fixes. Users urged rephrasing prompts to escape cycles.
- LM Studio 0.3.8 Boosts LaTeX and DeepSeek R1 “Thinking” UI: The update added LaTeX rendering and a DeepSeek R1 interface, fixing Windows installer bugs. Attendees lauded Vulkan GPU deduplication for stability.
- Perplexity’s Sonar Pro API Outguns Competitors in SimpleQA Benchmark: Sonar Pro dominated benchmarks with 73.3% math scores and 1M-token context, but rollout hiccups caused 500/403 errors. GDPR compliance debates flared for EU hosting.
Theme 5. Ethics, Censorship, and Workforce Displacement Fears
- DeepSeek R1’s 85% Performance Drop Fuels Censorship Suspicions: Users blamed R1’s overnight performance crash on internal censorship filters. Workarounds for sensitive prompts trended, while uptime monitors tracked fixes.
- Startup Founder Grapples with AI-Driven Layoff Guilt: A founder lamented potential job losses from their AI tool, sparking debates on automation ethics. Critics compared it to historical tech disruptions.
- Microsoft’s Phi-3.5 Safety Push Ignites Uncensored Fork: Phi-3.5’s “overzealous” censorship led to a Hugging Face fork. Coders criticized its refusal to answer benign queries, mocking its tic-tac-toe dodge.
PART 1: High level Discord summaries
Unsloth AI (Daniel Han) Discord
-
Dynamite 4-bit Quants from Unsloth: Unsloth introduced a dynamic 4-bit quantization method that avoids compressing select parameters, retaining strong accuracy at a fraction of typical VRAM usage.
- Users compared it with BnB 8-bit approaches, showing dynamic 4-bit preserves performance while only slightly raising memory demands.
- Phi-4 Fine-Tuning Fiasco: Users reported trouble with Phi-4 model's subpar output after fine-tuning, questioning if the model architecture is the culprit.
- They noted that fine-tuning worked on other notebooks, hinting that Phi-4 might need specialized settings for better responses.
- Chinchilla Crunch: Size vs. Tokens: Participants revisited the Chinchilla formula, spotlighting the interplay between model scale and training tokens for maximum efficiency.
- They observed how many models overshoot recommended thresholds, sharing empirical proofs of scaling gains that call for resource optimization.
- Synthetic Data Euphoria: Some argued for infinite synthetic data streams to sharpen model training, emphasizing eval compliance and curated inputs.
- They warned of garbage in, garbage out if data curation is haphazard, urging dedicated supervision for synthetic sets.
- Agentic AI 'All Hype' Claims: Members expressed frustration over marketing ploys labeling basic script-based systems as agentic AI, lacking true autonomous capacity.
- They pointed to the gap between grand brand statements and the straightforward reality of limited agent functionality.
Codeium (Windsurf) Discord
-
Web Search Woes: Codeium vs. Windsurf: One user pressed for Codeium to match Windsurf's web search capabilities, referencing the extension changelog.
- They questioned data handling policies by pointing to the Privacy Policy and comparing JetBrains' vs. VS Code features.
- Windsurf IDE Access Aggravations: Multiple members struggled to open FastAPI files on Ubuntu after the latest Windsurf update, noting that Cascade couldn't read or edit certain paths.
- A suggestion to adjust file permissions or environment settings emerged, but the issue persisted for some developers.
- Cascade’s Confusing Code Conversations: Several developers reported Cascade looping through repeated fixes when dealing with complicated code debugging prompts.
- They found that rephrasing instructions or giving more context helped end the cycle, showing the importance of well-structured prompts.
- Prompt Power & Project Memory: Users praised Windsurf's auto-generated memories for carrying context across sessions, referencing a Forum thread for deeper project memory ideas.
- They also cited a Cline Blog post emphasizing planning over perfect prompting to improve AI interactions.
- Diff Viewer Dilemma & Writemode Woes: Community members consistently faced color-scheme confusion in Windsurf's diff viewer and questioned if Writemode was free.
- A user clarified that it's a paid feature while others explored bridging multiple LLMs with Flow Actions for more flexible integrations.
LM Studio Discord
-
LM Studio 0.3.8 Bolsters Thinking UI: The newly released LM Studio 0.3.8 adds a Thinking UI for DeepSeek R1, plus LaTeX rendering via
\text{...}
blocks.- Attendees noted it addresses Windows installer issues and eliminates duplicate Vulkan GPUs, making usage more straightforward.
- DeepSeek R1 Distill Qwen 32B Wows Math Fans: Users praised DeepSeek R1 Distill Qwen 32B for outperforming other local models like Llama 405B on complex AIME-level math tests.
- They referenced the Hugging Face release for further details, praising its enhanced reasoning steps.
- Quantization Juggling on Apple Silicon: Participants explored 4-bit quantization to fit the R1 32B model on an M3 Max MacBook Pro, noting potential drops in answer quality.
- Multiple users tested MLX-optimized variants to cut memory demands, hinting at faster speeds despite possible precision trade-offs.
- High Price Tag for 671B Deployments: Estimates suggested $18,000 in hardware is needed for multiple NVIDIA Digits to run the full DeepSeek R1 671B, sparking debates about feasibility.
- Others mentioned a 4xA4000 setup as a cheaper route, remarking that bigger models do not always guarantee superior performance.
Nous Research AI Discord
-
Stargate's $500B Raises Eyebrows: Rumors spread that Project Stargate is eyeing a $500B investment from SoftBank, sparking skepticism given SoftBank's limited liquidity and debt as noted by Gavin Baker here.
- Discussion referenced Elon Musk's comment that SoftBank doesn't have the cash, fueling doubt about the proposal's validity.
- DeepSeek Doubts Over Razzberry: Members tested DeepSeek R1 Distill Llama-8B, observing how it externalizes reasoning on spelling 'razzberry' versus 'raspberry.'
- They highlighted comedic confusion about zero 'p's, illustrating DeepSeek's chain-of-thought reveal and its promise for deeper synergy among models.
- FLAME Blazes Through Excel: A 60M-parameter model called FLAME uses an Excel-specific tokenizer to tackle formula completion and repair, rivaling bigger models like Davinci.
- Community members admired its targeted training and smaller size, seeing it as a strong approach for domain-focused tasks.
- EvaByte Goes Token-Free: EvaByte debuts as a 6.5B byte-based model that uses 5x less data and achieves 2x faster decoding, welcoming multimodal possibilities without tokenizers.
- Skeptics questioned hardware efficiency, but results suggest a shift toward byte-oriented training with broader flexibility.
- STAR & TensorGrad Shake Up Model Architecture: STAR outlines an evolutionary approach for refining LLM structures, claiming performance gains in scaling and efficiency.
- TensorGrad introduces named edges for simpler matrix ops and symbolic optimization, attracting developers eager to discard tricky numeric dimension mapping.
Interconnects (Nathan Lambert) Discord
-
Stargate Scores Grandiose Growth: OpenAI's new Stargate Project secured a massive $500 billion plan over four years, with $100 billion immediately committed by SoftBank, Oracle, and MGX to fuel advanced AI infrastructure (reference).
- This colossal initiative is predicted to create numerous jobs and bolster American leadership in AI, with coverage appearing in Microsoft's official blog and various tech channels (blog link).
- Bespoke-Stratos-32B Battling Benchmarks: The Bespoke-Stratos-32B model, distilled from DeepSeek-R1, outperformed Sky-T1 and o1-preview in reasoning tasks, requiring 47x fewer training examples (source).
- It utilized an $800 open-source dataset to achieve cost-effective results, spurring interest in community-led data collection and collaborative improvements.
- Google Grants Another Billion for Anthropic: In a renewed show of confidence, Google infused $1 billion into Anthropic, fueling speculation on rolling funding strategies and AI competition (tweet).
- This investment reinforces Google's ongoing commitment to emerging AI players, with discussions highlighting possible expansions of Anthropic's next-gen models.
- Gemini-2.0-Flash Jumps to Arena Apex: Gemini-2.0-Flash-Thinking seized the #1 spot in the Chatbot Arena with a robust 73.3% math score and a 1 million token context window (link).
- Developers praised its potential for large-scale reasoning, while acknowledging that upcoming iterations could further refine its performance.
- GRPO Tweaks & T1 RL Triumph: Community members questioned the GRPO approach, pointing out concerns about KL divergence handling in advantage calculations, referencing TRL issues.
- Meanwhile, a new T1 paper from Ziphu and Tsinghua details scaled RL for large language models, blending trial-and-error with self-verification and cited in arXiv.
aider (Paul Gauthier) Discord
-
Gemini 2.0 Gains a Shiny Glow: Google introduced Gemini 2.0 with a 1 million token context window, 64K max output tokens, and native code execution support, as confirmed in this tweet.
- Users tested it against DeepSeek R1, expressing curiosity about how it handles complex tasks and overall efficiency.
- Aider’s Markdown Makeover: Users described a refined approach to Aider by storing major features, specs, and refactoring steps in markdown, referencing advanced model settings.
- They emphasized that LLM self-assessment paired with unit tests helps produce cleaner code and tighter development loops.
- Debating R1 vs. Sonnet: Community members observed distinctions between DeepSeek R1 and Sonnet, citing DeepSeek Reasoning Model docs for chain-of-thought capabilities.
- They noted Sonnet repeatedly offers more thorough suggestions and proposed merging R1's reasoning outputs into other models to tackle advanced scenarios.
- RAG with PDFs in Aider: A user asked about RAG for referencing PDFs in Aider, discovering that Sonnet supports it through a simple built-in command.
- That discussion inspired ideas about leveraging external data sources for deeper context in Aider’s workflows.
- Aider Upgrades & Nix Adventures: Multiple users hit upgrade errors and config roadblocks in Aider, swapping tips like removing the
.aider
directory or re-installing viapip
.
- They also pointed to a PR for Aider 0.72.1 in NixOS and pondered Neovim plugin setups, but final recommendations are still under review.
Stackblitz (Bolt.new) Discord
-
Bolt’s Big Bankroll Boost: Today, Bolt announced $105.5 million in Series B, led by Emergence & GV, with backing from Madrona and The Chainsmokers, as detailed in this tweet.
- They thanked the community for championing devtools and AI growth, promising to strengthen Bolt’s capabilities moving forward.
- Tetris Telegram Twist: A developer aims to build a Telegram mini app for Tetris with meat-themed bricks, sharing the Telegram Apps Center as a resource.
- They plan to incorporate a leaderboard, hoping this odd concept sparks community collaboration.
- Claude Conquers Code Confusion: Members found success using Claude to address tricky code tasks and retrieve policy updates when Bolt struggled.
- They noted Claude’s thoroughness in handling Supabase user permissions, praising the synergy of AI-driven debugging.
Yannick Kilcher Discord
-
DeepSeek R1 stirs competition with OpenAI: At 92% math performance, DeepSeek R1 outperforms Gemini and O1, gaining momentum for both technical prowess and cost advantages, as noted in DeepSeek R1 - API, Providers, Stats.
- Discussion highlights advanced geometric reasoning, multi-stage RL training, and a deep dive into DeepSeekMath (arxiv.org/abs/2402.03300) that pushes boundaries of model-based calculations.
- $500B Stargate Project sparks debate: OpenAI unveiled the Stargate Project with $500B in AI infrastructure funding, referencing their official tweet.
- Commentators questioned government-corporate overlap and U.S. technological gains, citing sources like AshutoshShrivastava.
- AI startup faces moral questions on job displacement: A founder shared remorse over potential layoffs triggered by their rapidly advancing AI solution, calling it a moral puzzle. Another user compared it to everyday automations that also cut positions, reflecting a larger socio-economic issue.
- Many agreed such side effects accompany new AI developments, emphasizing the tension between progress and workforce disruption.
- IntellAgent reframes agent evaluation: The open-source IntellAgent framework applies simulated interactions for multi-level agent diagnosis, showcased at GitHub.
- A corresponding paper at arxiv.org/pdf/2501.11067 shares surprising outcomes from data-driven critiques, with a visual workflow in
intellagent_system_overview.gif
. - UI-TARS and OpenAI Operator in the spotlight: Hugging Face released UI-TARS, targeting automated GUI tasks, including variants like UI-TARS-2B-SFT, as documented in their paper.
- Meanwhile, OpenAI readies an Operator feature for ChatGPT to execute browser actions, with reports from Stephanie Palazzolo.
OpenRouter (Alex Atallah) Discord
-
OpenRouter Slaps a Price Tag on Web Searches: The new $4/1k results pricing for web queries starts tomorrow, with an approximate cost of less than $0.02 per request.
- Members welcomed the streamlined model but noted curiosity about billing details once widespread usage begins.
- API Access Soft Launch Zooms In: OpenRouter’s API access arrives tomorrow with extra customizability options for users.
- Members anticipate testing these new features and sharing feedback on performance and integration.
- DeepSeek R1 Falters with 85% Performance Plunge: Reports indicated an 85% drop in DeepSeek R1’s API performance overnight, fueling worries about internal scrutiny and censorship filters.
- Some shared workarounds for sensitive prompts, while others monitored DeepSeek R1 – Uptime and Availability for official fixes.
- Cerebras Tantalizes with Mistral Large, Leaves Users Hanging: Many hoped to see Cerebras’ Mistral Large on OpenRouter, but it remains unavailable for public use.
- Frustrated users stick to Llama models instead, questioning whether Mistral Large is as ready as advertised.
Perplexity AI Discord
-
Sonar Pro Makes a Splash: Today, Perplexity rolled out the new Sonar and Sonar Pro API, enabling developers to incorporate generative search with real-time web research and strong citation features.
- Firms like Zoom tapped the API to elevate AI Companion 2.0, while Sonar Pro pricing touts lower costs than other offerings.
- SimpleQA Benchmark: Sonar Pro Steals the Show: Sonar Pro outperformed major rivals in the SimpleQA benchmark, outshining other search engines and LLMs in answer quality.
- Supporters praised web-wide coverage for being miles ahead of competing solutions.
- Comparisons Abound: Model Tweaks & Europe's Demands: Community members reported that Sonar Large now outruns Sonar Huge, with officials hinting at retiring the older model.
- Simultaneously, Europe's GDPR compliance push prompted discussion on hosting Sonar Pro in local data centers.
- Altman Teases 'PhD-Level Super Agents': A rumored briefing in D.C. by Altman mentioned advanced 'PhD-level Super Agents,' stirring curiosity about next-generation AI capabilities.
- Observers interpreted these hypothetical agents as a sign of major progress to come, though specifics remain sparse.
- Anduril's $1B Weapons Factory Gains Spotlight: News of Anduril establishing a $1B Autonomous Weapons Factory raised interest in defense-oriented machine systems, as shown in this video.
- Participants debated autonomous warfare and highlighted ethical questions tied to weaponized AI.
MCP (Glama) Discord
-
MCP's Code Crunch: Members discussed the evolving code editing feature in the MCP language server, citing improved capabilities for handling large codebases and synergy with Git operations.
- They mentioned a new Express-style API pull request that might unify how MCP server updates code in tandem with semantic tool selection.
- Brave Browsing for Crisp Docs: Users rely on Brave Search to fetch updated documentation and compile references into markdown for quick integration.
- They highlighted the brave-search MCP server approach for scraping and automating doc retrieval, praising the streamlined process.
- GPT's Grievances in 2024: Community members reported frustrations over custom GPT failing to incorporate new ChatGPT features, fueling doubts about the custom GPT marketplace.
- They noted concerns about these bots losing relevance, expressing disappointment in the slow pace of improvement.
- Claude Desktop's Prompt Parade: Participants explored hooking prompts into Claude Desktop, focusing on how to surface prompts via the prompts/List endpoint.
- They shared logging tool examples and partial code snippets, aiming to simplify the process for testing specialized prompts.
- Apify Actors & SSE Trials: Developers work on the MCP Server for Apify's Actors, building data extraction features but facing dynamic tool integration challenges.
- Issues around the Anthropic TS client underscore confusion over SSE endpoints, driving some members to pivot to Python while waiting on fixes.
Latent Space Discord
-
Ai2 ScholarQA Accelerates Literature Review: Ai2 ScholarQA launched a RAG-based solution for multi-paper queries, with an official blog post explaining cross-referencing capabilities.
- It helps academically-minded folks quickly gather research insights from various open-access papers, focusing on comparative analysis.
- Trump Launches $500B Project Stargate: President Trump announced Project Stargate, pledging a colossal $500B over four years to expand AI infrastructure in the US, with backing from OpenAI, SoftBank, and Oracle.
- Commentators like Elon Musk and Gavin Baker questioned the feasibility of that sum, calling it "ridiculous" but still acknowledging the ambition.
- Bespoke-Stratos-32B Emerges: A new reasoning model, Bespoke-Stratos-32B, distilled from DeepSeek-R1, showcases advanced math and code reasoning, per their announcement.
- Developers highlight that it uses "Berkeley NovaSky’s Sky-T1 recipe" to surpass previous models in reasoning benchmarks.
- Clay GTM Raises $40M: Clay GTM announced a $40M Series B expansion at a $1.25B valuation, with strong revenue growth that caught investor attention.
- Their existing funds remain largely untapped, and they plan to amplify momentum to drive further growth.
- LLM Paper Club Spotlights Physics of LMs: The LLM Paper Club event focuses on the Physics of Language Models and Retroinstruct, with details at this link.
- Attendees can add event alerts from Latent.Space by subscribing via the RSS logo, ensuring no event is missed.
GPU MODE Discord
-
NVIDIA's Blackwell Codegen Blazes On: The PR #12271 reveals Blackwell B100/B200 codegen 10.0 and teases an RTX 50 codegen 12.0, stirring anticipation for sm_100a and sm_101a.
- Despite the rumored sm_120 and a delayed Blackwell whitepaper, the community eagerly awaits more news and a talk on the Accel-Sim framework.
- Triton 3.2 HPC Hangups: Current TMA implementations can crash with
@triton.autotune
, causing confusion over persistent matmul descriptors and data dependency issues.
- A user pointed to the GridQuant gemm.py lines #79-100 for descriptor creation insights, underscoring the tricky nature of Triton kernels.
- GRPO Gains Momentum in RL: A minimal GRPO algorithm is nearly done, with its initial version set to run soon and early experiments already in progress.
- The Kimi-k1.5 paper and new GRPO trainer in TRL highlight growing interest in curriculum-based reinforcement learning.
- Accelerating LLM Inference with HPC Tools: A user sought faster Hugging Face
generate()
performance, referencing this commit and discussing liuhaotian/llava-v1.5-7b for heftier models.
- Meanwhile, this PyTorch post explores HPC-friendly strategies like specialized scheduling and memory optimizations to boost large model inference.
- Torch and Triton Tussle Over 3.2: The new Triton 3.2 disrupts
torchao
and torch.compile by droppingAttrsDescriptor
, recorded in PyTorch issue #144103.
- INT8 x INT8 dot product failures in Triton issue #5669 plus a major JIT refactor reveal repeated backward compatibility headaches.
Eleuther Discord
-
Google's Titan Tease & Transformers Tussle: After announcements that Google's Titans claim superior performance using advanced memory features, members discussed potential improvements in inference-time handling.
- Community sentiment suggested the original methods might be complicated to replicate, reflecting concern about sufficiently transparent experimentation.
- Grokking Gains with Numerics: A recent paper Grokking at the Edge of Numerical Stability highlighted how numerical issues hamper model training, fueling talk on improved optimization strategies.
- Members debated a first-order approach, referencing this GitHub repo for deeper insights.
- DeepSeek Reward Model Rumblings: Attendees examined the DeepSeek-R1-Distill-Qwen-1.5B model's metric differences, citing 0.834 (n=2) vs 97.3 (n=64) in partial evaluations.
- A link to the DeepSeek_R1.pdf referenced decoding strategies and architecture details for reward-based training.
- Minerva Math & MATH-500 Makeover: Participants tested minerva_math for symbolic equivalence with sympy, referencing the MATH-500 subset from OpenAI's 'Let's Think Step by Step' study.
- They debated whether DeepSeek R1 behaves like a base model or requires a chat template in these tasks, pointing to HuggingFaceH4/MATH-500 for further context.
- Linear Attention & Domino Learning: Members explored a specialized linear attention architecture aiming for speed without losing robust performance.
- They also discussed the Domino effect in skill stacking, referencing Physics of Skill Learning to highlight how sequential abilities emerge in neural networks.
OpenAI Discord
-
DeepSeek Duels with o1 & Sonnet: DeepSeek was compared with models like o1 and Sonnet in math and GitHub-related benchmarks, showing strong performance. It’s accessible for free on its official site and offers API integrations for varied platforms.
- Some users faced issues with o1, but they praised the official DeepSeek R1 features as more stable for quick evaluations, fueling a push toward consistent model alternatives.
- AI Security Sparks Curiosity: Members questioned why AI cybersecurity remains overshadowed, referencing how CrowdStrike has used ML for years. They see potential in generative AI for automated threat detection and code-based intrusion analysis.
- Community voices argued that corporations emphasize profits over fundamental security, pointing to a disconnect between marketing hype and genuine user protections.
- Image-Only GPT Gains Momentum: One user wanted to train a GPT-like model entirely on screenshots of chats, sidestepping text-based pipelines. They wondered if file uploads or image data handling is possible inside the existing chat completion API.
- Others weighed the feasibility of direct image ingestion, suggesting extra preprocessing steps until the API supports inline file support.
- OCR Confusions & Map Solutions: Members found OCR prompts caused extreme hallucinations, particularly with unconstrained examples. They explored a specialized workaround for reading maps, hoping OpenAI’s O series addresses spatial data soon.
- They warned about context contamination in free-form OCR setups, concluding domain-specific constraints are safer until better GIS support arrives.
Notebook LM Discord Discord
-
NotebookLM revolutionizes Church gatherings: One user leveraged NotebookLM to analyze 16 five-hour YouTube livestreams, generating a 250-page book and a 2000-page bible study.
- They referenced We Need to Talk About NotebookLM for motivation, applauding its capacity to handle overwhelming text volumes.
- Study workflows embrace NotebookLM: A user integrated NotebookLM in their study routine for weeks, crediting it with simplifying reference lookups.
- They shared a YouTube video highlighting its efficiency, motivating others to adopt a similar approach.
- Gemini gains momentum in prompt optimization: Members reported better NotebookLM outputs by pairing it with Gemini for refined instructions.
- They praised Gemini's impact on clarity but noted challenges in targeting highly specific documents.
- APA references and audio overviews spark debate: Participants wrestled with APA references, finding NotebookLM relies on previously used sources unless names are adjusted.
- They also discussed generating new audio overviews up to three times daily, warning about duplication and the need to remove old files first.
- CompTIA A+ content gains traction: A user unveiled part one of a CompTIA A+ audio series, with further installments on the way.
- Community members saw it as a key resource for self-paced certification prep, with NotebookLM lending quick information.
Stability.ai (Stable Diffusion) Discord
-
Flux Figures and Negative Nudges: One user reported that de-distilled flux models yield better performance on certain content with configured cfg settings, though they run more slowly.
- They noted that negative prompts can boost prompt adherence but add heavier compute overhead.
- AI Art Sparks Hostile Heat: Some participants encountered hostile responses when showcasing AI-generated art, including being told to kill myself for using these tools.
- They mentioned that negative sentiments toward AI art have lingered for years, prompting discussions on broader acceptance.
- Discord Bot Scams Bamboozle: Members flagged suspicious DMs from bot accounts asking for personal information, revealing a persistent scam trend.
- Someone recounted an earlier pitch for paid 'services,' suggesting these scams remain Discord's recurring problem.
- CitivAI Woes and Worries: A user highlighted CitivAI downtime multiple times daily, raising concerns about the service's stability.
- Others chipped in with similar experiences, questioning the platform's reliability.
- SwarmUI Face Fix Fanfare: One user asked about fixing faces in swarmUI, wondering if a refiner was needed to improve image fidelity.
- They noted the community's push for enhanced realism, aiming to refine image-generation pipelines further.
LlamaIndex Discord
-
AgentWorkflow Arrives & Amplifies Multi-Agent Systems: LlamaIndex announced AgentWorkflow, a new high-level framework that builds on LlamaIndex Workflows to support multi-agent solutions.
- They highlighted expanded tool support and community enthusiasm, calling it “the next step for more powerful agent coordination.”
- DeepSeek-R1 Dares & Defies OpenAI's o1: The DeepSeek-R1 model landed in LlamaIndex with performance comparable to OpenAI's o1 and works in a full-stack RAG chat app.
- Users praised its ease of integration and pinned hopes on “scaling it further for real-world usage.”
- Open-Source RAG System Gains Llama3 & TruLens Muscle: A detailed guide contributed step-by-step methods to build an open-source RAG system using LlamaIndex, Meta Llama 3, and TruLensML.
- It compared a basic RAG approach against an “agentic variant with @neo4j,” including performance insights on OpenAI vs Llama 3.2.
- AgentWorkflow Explores Parallel Agent Calls: Community members discussed parallel calls in AgentWorkflow while acknowledging agents run sequentially and tool calls can be asynchronous.
- They proposed “nesting workflows” as a possible hack to enable parallel tasks in multi-agent pipelines.
Modular (Mojo 🔥) Discord
-
Forum Feat Outshines Discord's Bounce: A user praised the forum for better archival and deeper discussion compared to the swift pace of Discord, referencing the general channel discussion.
- They suggested using it to avoid burying requests to Modular staff and maintain clarity on language design, encouraging consistent cross-posting of community showcases.
- Nightly Rolls In Quietly: There was a brief mention about Nightly being active and on the move, but no specifics were shared.
- The conversation teased it as an area of interest for future updates, leaving the community curious about new changes under the hood.
- Mojo Stays at Modular.com: Members wondered if Mojo would adopt a .org domain, reminiscent of Python, but the #mojo channel confirmed it remains under modular.com.
- They emphasized that Mojo won't split from Modular, aligning all efforts within the existing domain.
- MLIR Parallelization Gains Over LLVM: Users highlighted that MLIR parallelization is more promising than LLVM, referencing work in progress to bring parallel execution closer to reality.
- They see the new implementation as a key milestone for high-performance compilers, especially with LLVM still catching up.
- Rust Wrangles Work-Stealing Schedulers: The conversation tackled Rust's conflict between work-stealing schedulers and thread safety, noting it restricts mutex usage across yield points.
- While complicated, members advocated more granular control for tasks, suggesting that mindful concurrency design pays off despite the overhead.
LAION Discord
-
SunoMusic Soars with Sound: They introduced a feature letting users record themselves singing or playing instruments, showcased by this tweet.
- This new capability channels user creativity to produce entire songs from personal recordings.
- Audio Captioning Conundrum: Participants found it tricky to gauge background noise and recording quality due to limited datasets.
- They proposed adding synthetic noise to existing samples, referencing audio_augmentations.ipynb.
- Building Out Audio Datasets: One contributor maintains open sets covering voice acting, video clip embeddings, and scientific knowledge graphs.
- They face resource constraints but remain open to expansions from new collaborators.
- Emotional TTS Emerges in Bud-E: They teased an emotional Text-to-Speech roadmap for Bud-E, highlighting expressive voice output.
- A shared audio sample hints at deeper nuance in synthesizing user intent.
- Teacher Tackles AI Projects: A high school teacher manages classroom duties while expanding multiple AI dataset initiatives.
- They declined job offers to stay independent and rely on volunteer momentum for audio and video sets.
Cohere Discord
-
OpenAI Endpoint Explorations with DeepSeek: Members used the OpenAI endpoint with code examples referencing DeepSeek, showing how to define a custom route for chat requests, and highlighting flexible prompt structures.
- They emphasized the benefits of streamlined calls, praising the approach for enabling agile trial-and-error with various text generation frameworks, and expressed interest in further improvements.
- Best Model Buzz: 'Command-R7b' vs 'Aya Expanse 32B': A user requested the top text generation preferences, sparking mentions of Command-R7b and Aya Expanse 32B, with anecdotal experiences from multiple members.
- They pointed out distinct performance footprints, noting that certain use cases favored Command-R7b for heavier logic tasks, while others preferred Aya Expanse 32B for more expansive creative contexts.
- Cohere Command R+ 08-2024 Gains Traffic: Members highlighted the new Cohere Command R+ 08-2024 model in chat, praising its extended text generation under Azure AI Foundry setups.
- They discussed synergy with LlamaIndex workflows, expecting an eventual transition to Cohere API v2, and continued to share usage impressions.
- Meme Dreams: LCoT Model Weights for Cohere: Enthusiasts suggested Cohere release 'LCoT meme model weights' to merge comedic cues with deeper reasoning, referencing enterprise solutions for comedic expansions.
- Others questioned feasibility given Cohere’s brand focus but voiced hopes that it might yield specialized comedic text generation for new audiences.
- From Stills to Reels: Image-to-Video AI Plans: A user showcased attempts at image-to-video generation, referencing multiple ML frameworks, exploring cross-domain transformations, and seeking feedback.
- They singled out expansions toward 3D transitions and expressed excitement about fueling creative momentum in visual generative workflows.
LLM Agents (Berkeley MOOC) Discord
-
Syllabus Surprises by Jan 27: The revised syllabus for the upcoming MOOC is set to release by January 27, providing clarity on advanced LLM Agents content.
- Organizers note that final approvals are still underway, but learners can expect specific module details soon.
- Speakers Shaping The Stage: Suggestions to bring in Liang Wenfeng have been noted, though most guest speaker selections are already locked in.
- A new feedback form is planned to gather additional input from keen participants on future choices.
- Hackathon Hype on Hold: Currently, no official date is confirmed for the next event, and some are hopeful it could align with Spring plans.
- Organizers hinted that future research collaborations might dovetail with any forthcoming hackathon, so watchers should stay tuned.
- Spring MOOC Leaves Fall in the Dust: The new session will build on Fall content but won’t demand prior completion, letting anyone join.
- It expands on fundamental LLM Agent concepts, targeting both returning and fresh learners with updated materials.
- Research Collaborations Eyeing Future: Prof. Song is measuring interest in upcoming group research projects to showcase broader LLM Agent progress.
- Interested students are encouraged to mention their fields or topics, shaping any joint efforts that might emerge.
Nomic.ai (GPT4All) Discord
-
DeepSeek R1 Mystery with GPT4All: Community members asked which DeepSeek R1 distilled models run on GPT4All, noting a missing public model catalog comparable to LM Studio and hinting at llama.cpp updates as a requirement.
- They debated the timeline for these models’ release, exchanging tips on how to confirm readiness for a stable GPT4All integration.
- WordPress Chatbot Woes: A developer aimed to hook a chatbot into WordPress without a plugin and struggled to secure an API Key, raising concerns over official guidelines.
- Others chimed in with alternate routes, but the discussion ended with no concrete solution for an immediately accessible key.
- Quest for Free ChatGPT Access: A user openly sought free, unlimited use of ChatGPT, hoping for a viable workaround or generosity from the platform’s providers.
- The conversation highlighted a steady demand for no-cost AI solutions, with no consensus on legitimate free keys.
DSPy Discord
-
RAG Wrestles with Dynamic Data: A user asked how DSPy-based RAG manages dynamic data, highlighting interest in its real-time adaptability with no explicit solution provided.
- Others left the inquiry unresolved, giving no specific code examples or follow-up references.
- Call for DSPy Research Partners: A user requested collaboration on DSPy research, underscoring experience in AI for Good, LLMs, and Higher Education.
- They expressed a strong desire to make a meaningful contribution, though they provided no direct link or resource.
- LM Studio Snags in DSPy Integration: One user reported difficulty pairing LM Studio with DSPy, contrasting it with smoother experiences using Ollama.
- Another member asked 'How are you using the model?', prompting discussion about local environment setups and possible compatibility pitfalls.
- Ollama Error Ordeal: An error involving a 'str' object emerged when running DSPy with Ollama, hindering data-summary features.
- This forced DSPy to operate without a data-aware proposer, raising concerns about missing functionalities.
- Repo Spam Rant: A user complained about spam in a repository, possibly linked to a coin-related issue, calling it super lame.
- They saw it as an unwelcome distraction from real DSPy discussions, but no direct resolution was identified.
Torchtune Discord
-
Custom RLHF Overhaul Gains Support: Members proposed removing all custom losses while providing a doc page for easier additions, referencing issue #2206.
- They also highlighted the need for examples in passing custom forward functions, ensuring DPO full-finetuning compatibility.
- Phi 4 PR Waits for Final Tweaks: The Phi 4 PR drew attention for needing a few adjustments before merging, with the plan to refine its design soon.
- Contributors showed eagerness to address these issues swiftly, aligning with broader RLHF enhancements.
- SimPO Deprecation Rewrites the Roadmap: Developers declared SimPO deprecated to reduce clutter and push forward new RLHF recipes.
- They committed to updating related docs, aiming for more flexible loss integration in alignment tasks.
- Nature Communications Win Spurs Lively Cheers: The paper's feature in Nature Communications drew enthusiastic feedback from the community.
- Its acceptance underscored continuous research efforts, sparking super cool reactions and team pride.
OpenInterpreter Discord
-
OpenInterpreter 1.0 Pre-Release & Python Code Execution: The upcoming OpenInterpreter 1.0 sparked speculation about removing Python code execution, with members noting that it appears Python is not fully implemented in the 1.0 pre-release.
- One user asked if this feature might return later, reflecting uncertainty about future updates for coding functionality.
- Markdown & TXT Are Display Formats: A user clarified that Markdown and TXT files serve as text formatting mechanisms, rather than programming languages.
- A follow-up remark hinted that something might be in development to address formatting behaviors in OpenInterpreter.
Gorilla LLM (Berkeley Function Calling) Discord
-
Gorilla Model Gains Guidance: A member asked if the Gorilla model from the Ollama docs is correct, referencing the adrienbrault/gorilla-openfunctions-v2:Q6_K/model on Hugging Face.
- They tagged others to confirm correct references for building function-calling LLMs.
- LLaMA v2 Expands Capabilities: LLaMA v2 includes a 4096 context length, 30 block count, and 32 attention heads for greater capacity.
- Participants highlighted tokenizer settings and quantization version, stressing these finer points for advanced usage.
tinygrad (George Hotz) Discord
-
Tinygrad gets Windows treatment: A contributor prepared a pull request for Windows tests and asked if LLVM or Clang could power proper Windows backends.
- They included an image showing details of the PR, prompting questions on ensuring broad test coverage for multiple Windows setups.
- OpenCL GPU support takes priority: Another member stressed that GPU (OpenCL) support is the key addition for Windows testing, citing performance gains across different hardware.
- This discussion highlights the push to optimize Tinygrad for GPUs under Windows, refining cross-platform compatibility.
Axolotl AI Discord
-
Liger Kernel Gains KTO Cred: The Liger Kernel repository officially merged KTO loss, marking a significant update geared toward enhanced performance.
- Contributors celebrated this step, stressing that KTO loss could provide a bridge to better integration with upcoming model updates.
- Merging Models with the KTO Advantage: Engineers debated model merging strategies that might benefit from KTO-based metrics, citing newly incorporated features in Liger Kernel.
- They highlighted active collaboration, expecting this loss function to streamline merges in future releases.
Mozilla AI Discord
-
Local-First X AI Hackathon Launches in SF: Organizers announced the Local-First X AI Hackathon for Feb. 22 in San Francisco, highlighting offline-friendly development strategies.
- They expect a lively turnout, encouraging participants to brainstorm on advanced local-first AI before the event date.
- Join the Hackathon Conversation Thread: A dedicated discussion thread is now open for deeper discussions about project ideas and logistics.
- Organizers emphasize quick sign-ups and mention that new suggestions on local-first approaches remain welcome.
The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The HuggingFace Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
PART 2: Detailed by-Channel summaries and links
Unsloth AI (Daniel Han) ▷ #general (266 messages🔥🔥):
Model Quantization, Fine-tuning Models, DeepSeek R1 Model Support, Chat Templates and Thinking Tags, Dynamic 4-bit Quantization
-
Discussion on Model Quantization Techniques: Users debated the effectiveness of various quantization methods, with 8-bit quantization noted for its trade-off between performance and accuracy, while BnB 4-bit was discussed for its dynamic quantization capabilities.
- It was pointed out that while BnB 8-bit isn't well supported, alternative methods like FP8 and Q8_0 offer better flexibility and performance.
- Fine-tuning with Unsloth: A user inquired about fine-tuning the unsloth models with chat data, expressing that the model should ideally generate sequential messages rather than single responses.
- Suggestions included ensuring high-quality dataset examples and considering the employment of another model to supervise and review the outputs.
- DeepSeek R1 Compatibility: Clarifications about using the DeepSeek R1 model with llama-server revealed that users need to ensure they load the complete model rather than just a part of it for successful deployment.
- Discussion also indicated a broader support for running the model via VLLM, including large versions.
- Chat Templates and 'Thinking' Tags: Users explored the lack of a unified 'thinking' tag in chat templates for behavior training in DeepSeek, noting the resultant COT aspect in responses.
- Suggestions were made to either incorporate such templating or to avoid using specific templates entirely during training.
- Introduction of Dynamic 4-bit Quantization: The unsloth-bnb quantization method was introduced as a dynamic approach that aims to maintain accuracy while significantly reducing model size.
- It was emphasized that this method only slightly increases VRAM usage while effectively optimizing model performance.
Links mentioned:
- Unsloth - Dynamic 4-bit Quantization: Unsloth's Dynamic 4-bit Quants selectively avoids quantizing certain parameters. This greatly increases accuracy while maintaining similar VRAM use to BnB 4bit.
- Tweet from Fimbul (@fimbulvntr): Am I going crazy or is DeepSeek-R1 capped to a model_max_length of 16384?I think this is a bug. In reality it should be 163840.It has original_max_position_embeddings=4096 and a RoPE factor of 40... 4...
- Google Colab: no description found
- Google Colab: no description found
- tokenizer_config.json · deepseek-ai/DeepSeek-R1 at 3302ba78c0090838341caf8adfbe1e231308fa95: no description found
- So Cute Cat GIF - So Cute Cat Love - Discover & Share GIFs: Click to view the GIF
- facebook/layerskip-llama3-8B · Hugging Face: no description found
- bitsandbytes: no description found
- Datasets 101 | Unsloth Documentation: Learn all the essentials of creating a dataset for fine-tuning!
- unsloth/DeepSeek-R1-GGUF · Hugging Face: no description found
- GitHub - bagel-org/ZKLoRA: Efficient Zero-Knowledge Proofs for LoRA Verification: Efficient Zero-Knowledge Proofs for LoRA Verification - bagel-org/ZKLoRA
- tokenizer_config.json · unsloth/DeepSeek-R1-Distill-Llama-8B-unsloth-bnb-4bit at main: no description found
- 👨👨👧👧 GRPO by qgallouedec · Pull Request #2565 · huggingface/trl: What does this PR do?from datasets import load_datasetfrom peft import LoraConfigfrom trl import GRPOConfig, GRPOTrainer# Load the datasetdataset = load_dataset("trl-lib/tldr", spli....
Unsloth AI (Daniel Han) ▷ #off-topic (6 messages):
Unsloth training on Medium, Weights & Biases ETA tracking, Custom code for ETA, Fine-tuning challenges
-
Unsloth's Fine-tuning Explored on Medium: A Medium article discussed fine-tuning LLMs using Unsloth with Weights & Biases and vLLM for model serving, addressing the need for adaptation to specific tasks.
- It highlights challenges in LLM fine-tuning, especially concerning GPU memory and computation time.
- Weights & Biases Lacks ETA Tracking: A user expressed disappointment that Weights & Biases does not show an estimated time of arrival (ETA) for training processes, only displaying run time and start time.
- They received confirmation that Weights & Biases doesn’t track time but instead focuses on plotting training metrics.
- Custom Code Needed for ETA: A community member clarified that while Weights & Biases does not have built-in ETA tracking, it can be done by writing custom code.
- This provides a workaround for users wanting to monitor training completion times more precisely.
Link mentioned: Fine-Tuning Llama-3.1-8B for Function Calling using LoRA: Leveraging Unsloth for fine-tuning with Weights & Biases integration for monitoring and vLLM for model serving
Unsloth AI (Daniel Han) ▷ #help (146 messages🔥🔥):
Phi-4 model issues, Running DPO training script, Unsloth notebooks updates, Using Triton with Unsloth, Fine-tuning suggestions for different models
-
Challenges with Phi-4 Model Fine-Tuning: Users reported difficulties in fine-tuning the Phi-4 model, specifically regarding continuous generation of poor responses and confusion over output options.
- It was suggested that these issues could be model-specific, as other notebooks worked as expected for different models.
- DPO Training Script Import Errors: A user encountered an ImportError when running the DPO training script due to relative imports in the Triton library and was advised to use absolute import instead.
- Another user experienced a RuntimeError related to compiling llama.cpp and was advised to check for updates and dependencies.
- Updates and Fixes to Unsloth Notebooks: Multiple users commented on the recent updates to Unsloth notebooks, including fixes for errors related to module imports and memory management issues.
- It was confirmed that the notebooks are designed to be reproducible, focusing on specific models to avoid unrelated errors during training.
- Fine-Tuning on CPU Challenge: Users discussed the difficulties of fine-tuning on CPU due to slow performance and memory constraints, especially with larger models.
- Suggestions included lowering batch sizes or using specialized templates for specific models to optimize resource use.
- Adjustments for Successful Fine-Tuning: Users shared insights on optimizing settings for LoRA and batch sizes to improve fine-tuning performance across different models.
- Recommendations included ensuring correct dataset formatting and using established benchmarks or templates to guide adaptations.
Links mentioned:
- Google Colab: no description found
- Google Colab: no description found
- Quazim0t0/Phi4.React.Turn.V2.Full · Hugging Face: no description found
- GitHub · Build and ship software on a single, collaborative platform: Join the world's most widely adopted, AI-powered developer platform where millions of developers, businesses, and the largest open source community build software that advances humanity.
- Quantization: no description found
- Unsloth Notebooks | Unsloth Documentation: Below is a list of all our notebooks:
- unsloth/unsloth/kernels/fast_lora.py at d802bbf4e298cb0da1e976ab9670fbc1cbe3514c · unslothai/unsloth): Finetune Llama 3.3, Mistral, Phi-4, Qwen 2.5 & Gemma LLMs 2-5x faster with 70% less memory - unslothai/unsloth
- trl/trl/scripts/dpo.py at a9b54a852ee12ff508773edb02e1c243817e71ae · huggingface/trl: Train transformer language models with reinforcement learning. - huggingface/trl
- unsloth/unsloth/models/dpo.py at main · unslothai/unsloth: Finetune Llama 3.3, Mistral, Phi-4, Qwen 2.5 & Gemma LLMs 2-5x faster with 70% less memory - unslothai/unsloth
- GitHub - ggerganov/llama.cpp: LLM inference in C/C++: LLM inference in C/C++. Contribute to ggerganov/llama.cpp development by creating an account on GitHub.
Unsloth AI (Daniel Han) ▷ #research (191 messages🔥🔥):
Synthetic Data Training, Chinchilla Optimal Models, Emotion Tracking in AI, Agentic AI Claims, Dynamic vs. Static Learning Systems
-
Synthetic Data for Training: A discussion unfolded on the potential of training AI using infinite synthetic data streams, suggesting that this could optimize training efficiency by fine-tuning around eval compliance.
- The argument noted that proper data curation is essential to avoid 'garbage in, garbage out' when utilizing synthetic datasets.
- Understanding Chinchilla Optimality: Participants discussed Chinchilla models, which aim to find the right balance of model size and training tokens to optimize performance.
- It was highlighted that existing models are far beyond optimal thresholds, with empirical evidence showing massive scaling efficiency gains in modern architecture.
- Challenges with Emotion Tracking: One member recounted experiences in developing bots for the erotic industry, emphasizing the importance of emotion propagation and psychological principles in bot responses.
- The conversation touched on the complexities of implementing emotional tracking effectively due to the dynamic nature of human interactions.
- Skepticism Towards Agentic AI: There was a consensus that many claims about autonomous agents are exaggerated and often equate to outdated concepts being marketed as new.
- Members expressed frustration with the industry selling the idea of agentic AI while only delivering basic scripting capabilities.
- Potential of AI in Education: The group contemplated the future of AI, envisioning possibilities of large language models teaching elementary school children, particularly in math, through interactive mediums.
- Concerns were raised about whether such learning could be effectively implemented within current frameworks, without overpromising on AI capabilities.
Links mentioned:
- TinyStories: How Small Can Language Models Be and Still Speak Coherent English?: Language models (LMs) are powerful tools for natural language processing, but they often struggle to produce coherent and fluent text when they are small. Models with around 125M parameters such as GP...
- Papers with Code - Chinchilla Explained: Chinchilla is a 70B parameters model trained as a compute-optimal model with 1.4 trillion tokens. Findings suggest that these types of models are trained optimally by equally scaling both model size a...
Codeium (Windsurf) ▷ #discussion (48 messages🔥):
Codeium Extension Updates, Windsurf IDE Issues, Model API Integration, Diff Viewer Difficulties, Privacy Policy Queries
-
Codeium Extension Lacks Web Search: A user inquired about when the Codeium extension would receive web search capabilities similar to Windsurf, noting the current limitations in features.
- Another member expressed that many JetBrains IDE users find VS Code's UX insufficient despite its advantages.
- Access Issues in Windsurf IDE: Several users reported being unable to access FastAPI files such as main.py and server.py after a recent upgrade in the Windsurf IDE.
- One user mentioned that Cascade coding assistant could no longer access these files on a Linux Ubuntu system.
- Interest in Custom API Integrations: A user suggested that allowing developers to use their own APIs combined with Flow Actions would enhance the platform.
- Another participant added that integrating any LLM via a standard API could be a beneficial option alongside current subscriptions.
- Writemode Feature Pricing: A member asked if the writemode feature is free, to which another clarified it is a paid feature.
- The clarification highlights the distinction between free and premium functionalities within the platform.
- Diff Viewer Confusion: A user expressed difficulty understanding the diff viewer's color scheme and sought ways to change it.
- Additionally, they noted issues with the cursor not changing in insert mode while using VIM with Windsurf.
Links mentioned:
-
Changelog | Visual Studio Marketplace
: no description found - Privacy Policy | Windsurf Editor and Codeium extensions: Codeium is the AI code assistant platform that developers love and enterprises trust. Also the builders of Windsurf, the first agentic IDE. - Warp: The intelligent terminal: Warp is the intelligent terminal with AI and your dev team's knowledge built-in. Available now on MacOS and Linux. - JetBrains meets Codeium Chat — level up your coding: Codeium Chat is now available for FREE on JetBrains IDEs such as IntelliJ, WebStorm, PyCharm, AndroidStudio, and more! Codeium users can use the AI assistant...
Codeium (Windsurf) ▷ #windsurf (426 messages🔥🔥🔥):
Windsurf's Auto-Generated Memories, Development Challenges with Cascade, Prompt Engineering and Context, AI Integration in Programming, User Experiences with Windsurf
-
Windsurf's Auto-Generated Memories Feature: Users expressed excitement about Windsurf's auto-generated memories feature, which seems to capture important conversation context within projects.
- However, there are concerns about how these memories are structured and their effectiveness when switching between different workspaces.
- Challenges in Development with Cascade: Several users mentioned experiencing issues with Cascade, citing problems like looping prompts and repeated fix attempts leading to frustration.
- Engaging Cascade in discussions about specific code issues often helps resolve these looping scenarios.
- The Importance of Prompt Engineering and Context in AI: It was noted that prompting techniques significantly influence the responses received from AI, emphasizing the need for well-structured requests.
- Discussion participants suggested that improving the AI's memory capabilities could alleviate common issues faced by developers when interacting with it.
- Future AI Integration in User Workflow: Participants envisioned a future where AI tools like Windsurf could provide more seamless context integration across various applications and devices.
- This ideal scenario would enhance productivity and reduce the reliance on prompt engineering and context management.
- User Experiences with Windsurf and Coding Practices: Users shared their experiences with Windsurf, including its potential to enhance coding efficiency and automate tasks like Git history checks.
- Concerns about the stability of the tool and the performance of features like auto-complete were also raised.
Links mentioned:
- The 70% problem: Hard truths about AI-assisted coding: A field guide and why we need to rethink our expectations
- Web Search - Codeium Docs: no description found
- D-Wave Leap Log In | D-Wave Leap™: no description found
- Support | Windsurf Editor and Codeium extensions: Need help? Contact our support team for personalized assistance.
- Why AI Engineers Need Planning More Than Perfect Prompts - Cline Blog: no description found
- Post Not Found: no description found
- Windsurf forked VS Code to compete with Cursor. Talking the future of AI + Coding: Wes and Scott talk with Kevin Hou and Varun Mohan from Windsurf about the evolving landscape of AI in coding, and the future of software development.👉 Join ...
- Web Search Best Practices: Save Credits and Optimize Your Workflow - Windsurf Editor: Ready to get the most out of Windsurf's brand-new Web Search feature? This deep dive is here to help you unlock its full potential!In this video, you’ll lear...
- Persistent, intelligent project memory: .cursorrules is a stopgap. What we really need is for Cursor to truly remember interactions with the user and what the project needs and auto-update this memory as the user interacts with Cursor and ...
- GitHub - kinopeee/windsurfrules: Contribute to kinopeee/windsurfrules development by creating an account on GitHub.
LM Studio ▷ #announcements (1 messages):
LM Studio 0.3.8, Thinking UI, LaTeX rendering, Bug fixes
-
LM Studio 0.3.8 Launches with Exciting Features: The latest version, LM Studio 0.3.8, introduces a Thinking UI for DeepSeek R1 and new support for rendering LaTeX in
ext{...}
blocks.- Get it through an in-app update or download it from the website; a detailed demo showcases the enhancements.
- Enhanced LaTeX Rendering in LM Studio: With this update, users can now render math expressions effectively with LaTeX, enhancing the usability for technical content creation.
- The math can be wrapped in
egin{equation}... ext{}
or inline withext{...}
, streamlining mathematical communication. - A Series of Bug Fixes for Smooth Operation: Version 0.3.8 addresses several bugs, including incorrect bundling of LM Runtimes in the Windows installer and duplicate Vulkan GPUs being displayed.
- Additionally, it resolves an issue where messages in old chats would not show, ensuring a more reliable user experience.
Link mentioned: Tweet from LM Studio (@lmstudio): LM Studio 0.3.8 🚢- Thinking UI for DeepSeek R1- LaTeX rendering improvements- Bug fixes
LM Studio ▷ #general (209 messages🔥🔥):
DeepSeek R1 performance, Model loading issues on Mac, Quantization settings, Using LMStudio effectively, Math problem-solving capabilities of models
-
DeepSeek R1 excels in math problems: The R1 Distill Qwen 32B model is highlighted as the best local model for complex contest math problems, possibly competing with top AIME exam takers.
- Users noted its superior performance compared to Llama 405B and DeepSeek V3 671B.
- Model loading difficulties on Mac M1 Max: One user reported difficulty loading the DeepSeek 32B model on a MacBook Pro M1 Max, encountering an error related to the model vocabulary.
- They sought help with accessing runtimes and updating llama.cpp to resolve the issue.
- Confusion over LMStudio settings: Inquiries were made about the 'Keep Model in Memory' setting in LMStudio on Apple Silicon, with mixed results on RAM usage.
- It was suggested that these settings might not make a difference due to unified memory architecture.
- Quantization performance discussions: Users discussed how quantization settings, especially Q4, impact model performance and memory usage.
- Different models were recommended based on available RAM and desired inference speed.
- Enhancements in AI learning and usage: Several users expressed satisfaction with the progress made using LMStudio, particularly in terms of AI learning and answering specific queries.
- Excitement was noted about future developments that could make AI models even more lightweight.
Links mentioned:
- Tweet from Yorkie (@heyitsyorkie): @levelsio @levelsio you need to update your LM Studio version: https://lmstudio.ai/0.2.31 is old, there's been a complete overhaul in 0.3.*
- Tweet from thebes (@voooooogel): Made a very stupid sampler-based lever to try and mimic o1-style "reasoning_effort=high" on r1:If appears before enough thinking tokens have been generated, the sampler replaces...
- GitHub - deepseek-ai/DeepSeek-R1: Contribute to deepseek-ai/DeepSeek-R1 development by creating an account on GitHub.
LM Studio ▷ #hardware-discussion (180 messages🔥🔥):
DeepSeek R1 Pricing, MacBook Performance for LLMs, GPUQuantization, NVIDIA Digits Requirements, Model Size Efficiency
-
Understanding the Price of Running DeepSeek R1 Models: The requirement for multiple NVIDIA Digits to run the full DeepSeek R1 671B model means a hefty investment, with some estimating costs around $18,000 for the required hardware.
- Conversely, options like 4xA4000 can provide an efficient solution for running models while managing cost, suggesting innovative alternatives.
- MacBook Air vs Pro for LLMs: Discussions highlight the M4 MacBook Air's memory bandwidth versus the M2 MacBook Air, weighing whether it's a worthy upgrade for LLM functionalities.
- Concerns arose about the Air's ability to handle thermal limits compared to the more robust MacBook Pro models.
- GPU Quantization for Model Fitting: Users noted that running the R1 32B model on an M3 Max MacBook Pro may be challenging without using 4-bit quantization, which could compromise quality.
- Tips shared included using MLX versions optimized for Apple silicon to enhance performance while managing memory needs.
- Managing GPU VRAM for Optimal Performance: A discussion emerged about running models with higher VRAM requirements, stressing that using the same card for inference could lead to VRAM contention during operation.
- It was emphasized that for optimal configurations, separating inference workloads from displayed graphics is advisable, particularly on Windows systems.
- Critique of Large-Scale Models: Participants debated the practicality of training and running extremely large models like DeepSeek R1, suggesting more compact, efficient models may yield better results.
- Concerns were raised about efficiency and the reality that larger models do not inherently guarantee improved performance, encouraging a focus on targeted, high-quality models.
Link mentioned: bartowski/DeepSeek-R1-Distill-Qwen-32B-GGUF · Hugging Face: no description found
Nous Research AI ▷ #general (295 messages🔥🔥):
AI Model Development, Chip Manufacturing Competition, Blockchain and Crypto, DeepSeek Reasoning Extraction, Project Stargate Funding
-
Project Stargate's Ambiguous Funding: Discussion centered around the proposed $500 billion funding for Project Stargate, with skepticism about its feasibility given SoftBank's financial situation and capital limitations.
- Analysts questioned the viability of such a massive investment in the absence of significant changes in SoftBank's equity and cash flow.
- Chip Manufacturing Landscape: The conversation highlighted the challenges in chip manufacturing, particularly the reliance on TSMC and capacity issues among competitors like Samsung and Intel.
- Participants expressed concern about whether alternative chip manufacturers could effectively compete without significant investment and time.
- Implications of Blockchain and Crypto: Participants debated the potential of decentralized financial systems, with an emphasis on how increasing control by governments could fuel interest in cryptocurrencies.
- Despite potential benefits, concerns were raised about the practicality of crypto adoption in face of regulatory pressures and government capability to ban it.
- Insights on DeepSeek Reasoning Extraction: DeepSeek's capability to extract reasoning processes was discussed, with an emphasis on its potential benefits for models like Claude and O1.
- There's intrigue about using reasoning extraction not just for cost savings but also for integrating deeper reasoning capabilities into other models.
- Future of AI Development: Participants speculated on the future evolution of AI models, particularly the anticipated improvements in DeepSeek v3 trained on R1 outputs.
- The conversation pointed toward hopes for better performance as reasoning processes from various models are integrated and refined.
Links mentioned:
- Bloomberg - Are you a robot?: no description found
- Tweet from Ivanka Trump (@IvankaTrump): Leopold Aschenbrenner’s SITUATIONAL AWARENESS predicts we are on course for Artificial General Intelligence (AGI) by 2027, followed by superintelligence shortly thereafter, posing transformative oppor...
- Tweet from Demis Hassabis (@demishassabis): Our latest update to our Gemini 2.0 Flash Thinking model (available here: https://goo.gle/4jsCqZC) scores 73.3% on AIME (math) & 74.2% on GPQA Diamond (science) benchmarks. Thanks for all your feedbac...
- How blockchains could change the world: Ignore Bitcoin’s challenges. In this interview, Don Tapscott explains why blockchains, the technology underpinning the cryptocurrency, have the potential to revolutionize the world economy.
- Joseph717171/DeepSeek-R1-Distill-Llama-8B-OQ8_0-F32.EF32.IQ4_K-Q8_0-GGUF · Hugging Face: no description found
- Tweet from Sun 乌龟 💖 (@suntzoogway): Guys this is a parody I wrote up!Just trying to hyperstition a good future (I'm trapped in EU regulatory hell)
- Tweet from Teknium (e/λ) (@Teknium1): I am seeking two engineers for the post training team at Nous Research to build the future of generally capable models, explore cognizant, creative models, and advance state of the art reasoning and a...
- Tweet from Pietro Schirano (@skirano): By the way, you can extract JUST the reasoning from deepseek-reasoner, which means you can send that thinking process to any model you want before they answer you. Like here where I turn gpt-3.5 turbo...
- Tweet from Gavin Baker (@GavinSBaker): Stargate is a great name but the $500b is a ridiculous number and no one should take it seriously unless SoftBank is going to sell all of their BABA and ARM.SoftBank has $38b in cash, $142b in debt an...
- Tweet from Elon Musk (@elonmusk): @OpenAI They don’t actually have the money
Nous Research AI ▷ #ask-about-llms (6 messages):
Mechanical Interpretation of Model Activations, DeepSeek Capabilities, Synthetic Data Generation
-
Members Explore Mechanical Interpretation Techniques: A member is seeking advice on how to visualize the evolution of model activations through layers given large data inputs in their hobby project on mechanical interpretation.
- Another member mentioned knowing someone who has experience with this, promising to refer back after checking old messages.
- DeepSeek Experiment Shows Variability in Responses: Experimentation with DeepSeek yielded interesting results where it calculated responses correctly but questioned its own conclusions regarding spelling variants of 'razzberry'.
- The output indicated zero p's in 'razzberry' but pointed out potential confusion with 'raspberry', demonstrating the tool's internal reasoning process.
- Inquiry on Synthetic Data Generation Guidelines: A member is inquiring about resources that outline the do's and don'ts of generating synthetic data for finetuning models.
- This reflects ongoing interest in best practices for synthetic data usage among the community.
Link mentioned: Tweet from LLM Fan (@llm_fan): I think LLM's are catching onto the 'r's in strawberry question. I asked deepseek"how many p's are there in the correct word for razzberry?" ( I tried 'proper word' als...
Nous Research AI ▷ #research-papers (5 messages):
FLAME model, Human-AI representation alignment
-
FLAME: The Small and Mighty Model: The paper introduces FLAME, a transformer-based model specifically trained on Excel formulas, achieving competitive performance with only 60M parameters as opposed to larger models like Davinci.
- This model utilizes a clever training dataset curation through sketch deduplication and introduces an Excel-specific formula tokenizer.
- Human and AI Representation Convergence: A recent study reveals that both high-performing ANNs and biological brains converge on similar representations, suggesting a universal structural alignment.
- By identifying stimuli that vary inter-model representation agreement, the research provides insight into how ANNs can map biological representations and computations.
Links mentioned:
- FLAME: A small language model for spreadsheet formulas: Spreadsheets are a vital tool for end-user data management. Using large language models for formula authoring assistance in these environments can be difficult, as these models are expensive to train ...
- Universality of representation in biological and artificial neural networks: no description found
Nous Research AI ▷ #interesting-links (12 messages🔥):
Automated architecture search for LLMs, EvaByte tokenizer-free model, Tensor networks in ML
-
Automated Architecture Search Sparks Curiosity: A paper on automated architecture search for LLMs discusses a novel approach that combines hierarchical numerical encoding with evolutionary algorithms to optimize model architectures.
- While some doubt its uniqueness as a competitive advantage, the findings indicate significant advances in optimizing deep learning frameworks.
- Meet EvaByte: The Tokenizer-Free Wonder: EvaByte, a collaboration between the University of Hong Kong and SambaNova Systems, introduces a 6.5B byte-level language model that rivals modern tokenizer-based LMs, utilizing 5x less data and achieving 2x faster decoding.
- It operates entirely on a per-byte basis, offering flexibility and improved performance on various tasks, despite concerns regarding irregular structures affecting hardware utilization.
- Tensor Networks Make Waves in ML: A new ML library called TensorGrad utilizes named edges instead of numbered indices, simplifying operations like 2D convolution while performing symbolic reasoning and optimization.
- With features like common subexpression elimination between passes, it generates efficient compiled torch code and requires user feedback to enhance ease of use.
Links mentioned:
- EvaByte: Efficient Byte-level Language Models at Scale | HKU NLP Group : no description found
- STAR: Synthesis of Tailored Architectures: Iterative improvement of model architectures is fundamental to deep learning: Transformers first enabled scaling, and recent advances in model hybridization have pushed the quality-efficiency frontier...
- Tweet from Lin Zheng (@linzhengisme): 🚀 Meet EvaByte: The best open-source tokenizer-free language model! Our 6.5B byte LM matches modern tokenizer-based LMs with 5x less data & 2x faster decoding, naturally extending to multimodal tasks...
- GitHub - thomasahle/tensorgrad: Tensor Network Library with Autograd: Tensor Network Library with Autograd. Contribute to thomasahle/tensorgrad development by creating an account on GitHub.
- evabyte/evabyte_hf/eva.py at ba8f65c5fe502b7ed07f916773754734b91b52fd · OpenEvaByte/evabyte: EvaByte: Efficient Byte-level Language Models at Scale - OpenEvaByte/evabyte
Nous Research AI ▷ #research-papers (5 messages):
FLAME model, Human-AI representation similarities, Small model advantages
-
FLAME: The Smol Model for Excel Formulas: The paper presents FLAME, a transformer-based model trained exclusively on Excel formulas that contains 60M parameters, achieving competitive performance with significantly less data.
- Using Excel-specific tokenization and pre-training objectives, FLAME outperforms larger models like Davinci on tasks such as formula repair and completion.
- Understanding Human-AI Representation Alignment: A study reveals that high-performing ANNs and brains converge onto the same representations, highlighting the alignment between model behavior and biological systems.
- By identifying stimuli that vary inter-model representation agreement, the research provides insights into the features that distinguish high-agreement sentences and images.
- Affection for Small Models: Community discussions reveal enthusiasm for small models, with members sharing their appreciation for models like FLAME.
- The trend suggests a growing interest in models that are efficient yet powerful, aligning with recent advancements in AI research.
Links mentioned:
- FLAME: A small language model for spreadsheet formulas: Spreadsheets are a vital tool for end-user data management. Using large language models for formula authoring assistance in these environments can be difficult, as these models are expensive to train ...
- Universality of representation in biological and artificial neural networks: no description found
Nous Research AI ▷ #reasoning-tasks (1 messages):
lowiqgenai: Hey i did some using MistralAI free Services fhai50032/medmcqa-solved-thinking-o1
Interconnects (Nathan Lambert) ▷ #news (87 messages🔥🔥):
Stargate Project Funding, Updates on AI Models, Google's Investment in Anthropic, Flash Thinking Model, Concerns over AI Data Usage
-
Stargate Project Launches with Massive Funding: OpenAI has announced the Stargate Project, aiming to invest $500 billion over four years to build new AI infrastructure, starting with $100 billion immediately to drive American leadership in AI.
- The project has initial funding from SoftBank, Oracle, and MGX, creating an estimated hundreds of thousands of jobs while enhancing national security.
- Bespoke-Stratos-32B Model Outperforms Competitors: The new Bespoke-Stratos-32B model, distilled from DeepSeek-R1, has surpassed both Sky-T1 and o1-preview in reasoning benchmarks while requiring 47 times fewer examples for training.
- The dataset used in its training cost $800 and is available in open-source form to foster collaborative development.
- Google's Commitment to Anthropic: Google has made headlines by investing another $1 billion into Anthropic, highlighting a rolling funding strategy that raises questions about the company's funding approach.
- This investment continues Google's support of newer AI technologies amidst concerns about competition in the AI landscape.
- Flash Thinking Model Rises to Number One: Gemini-2.0-Flash-Thinking has claimed the top spot in the Chatbot Arena, showing significant performance boosts across various domains.
- The model achieved a score of 73.3% on mathematics and features a 1 million token content window, promising improvements in next iterations.
- Concerns Over AI Data Usage Policies: Concerns were raised over data usage policies of AI services, particularly regarding whether user inputs in free services would be used for training purposes.
- A clarification from Google indicated that both paid and free services are subject to similar data usage policies upon activation of a Cloud Billing account.
Links mentioned:
- Tweet from undefined: no description found
- Microsoft and OpenAI evolve partnership to drive the next phase of AI - The Official Microsoft Blog: We are thrilled to continue our strategic partnership with OpenAI and to partner on Stargate. Today’s announcement is complementary to what our two companies have been working on together since 2019. ...
- The Stargate Initiative: Microsoft and OpenAI's $100 Billion Data Center Project: Microsoft and OpenAI are setting a new benchmark in the field of AI.
- Tweet from Logan Kilpatrick (@OfficialLoganK): @HCSolakoglu GA in Jan for 2.0 flash (non thinking version)
- Tweet from Logan Kilpatrick (@OfficialLoganK): RE: "Gemini is frustrating in that they very clearly state they won't train on your input to their paid models... but both AI Studio and their Gemini preview models are free.."We just upda...
- Tweet from Logan Kilpatrick (@OfficialLoganK): RE: "Gemini is frustrating in that they very clearly state they won't train on your input to their paid models... but both AI Studio and their Gemini preview models are free.."We just upda...
- Tweet from Noam Shazeer (@NoamShazeer)): Your feedback on Gemini 2.0 Flash Thinking has been incredible—thank you!We’ve taken your suggestions and made an experimental update…
- Tweet from lmarena.ai (formerly lmsys.org) (@lmarena_ai): New Gemini-2.0-Flash-Thinking is now #1 in Chatbot Arena⚡🤔Highlights:- Scores highest, overtaking Gemini-Exp-1206- +17 pts boost over the previous 1219 checkpoint- #1 across all domains (hard, coding...
- Tweet from adi (@adonis_singh): anthropic are deprecating claude 3 sonnet. could be because they plan on releasing 4 sonnet soon..
- Tweet from Agus 🔎 🔸 (@austinc3301): Ah, yes. Of course we’re naming this project after the fictitious portal by which several hostile alien civilizations tried to invade and destroy Earth.Quoting OpenAI (@OpenAI) Announcing The Stargate...
- Tweet from NIK (@ns123abc): BREAKING: Google invests another $1 billion in Anthropic
- Tweet from Mahesh Sathiamoorthy (@madiator): Introducing Bespoke-Stratos-32B, our reasoning model distilled from DeepSeek-R1 using Berkeley NovaSky’s Sky-T1 recipe. The model outperforms Sky-T1 and o1-preview in reasoning (Math and Code) benchma...
- Tweet from Prime Intellect (@PrimeIntellect): Today, we are releasing:- INTELLECT-MATH, a frontier 7B parameter model for math reasoning- The largest synthetic math dataset to date of 5M verified reasoning traces- An outlook on decentralized trai...
- Tweet from Noam Brown (@polynoamial): This is on the scale of the Apollo Program and Manhattan Project when measured as a fraction of GDP. This kind of investment only happens when the science is carefully vetted and people believe it wil...
- Tweet from Tibor Blaho (@btibor91): According to The Information's report, OpenAI's Operator will provide in-ChatGPT browser automation with user controls, task sharing, and login persistence, except for Gmail- OpenAI's Oper...
- Tweet from Sam Altman (@sama): watching @potus more carefully recently has really changed my perspective on him (i wish i had done more of my own thinking and definitely fell in the npc trap).i'm not going to agree with him on ...
- Tweet from Sam Altman (@sama): build monuments in the desert
- 豆包大模型1.5Pro正式发布: no description found
Interconnects (Nathan Lambert) ▷ #ml-drama (16 messages🔥):
Microsoft's Investment in Stargate, Billionaires on Twitter, AI Safety Discussions, Model Alignment, Influence of Tech Leaders
-
Microsoft vows support for Project Stargate: In response to Elon Musk's concerns, Microsoft CEO stated, 'All I know is, I’m good for my $80 Billion' regarding funding for Project Stargate, sparking a flurry of reactions.
- This comes amid assumptions that Microsoft would invest significantly in data centers this year, fueling further speculation on financial backing.
- Billionaires banter on Twitter: A member remarked about the entertainment value of billionaires engaging in 'battles for likes on Twitter', highlighting the absurdities of tech mogul rivalries.
- The discussion included various witty remarks about CEOs and their public exchanges, further emphasizing the comical nature of these interactions.
- Concerns over AI Safety Metrics: A post questioned the absence of information on alignment and safety practices for the MiniCPM model, suggesting it needs robust assessment metrics like the FLI AI Safety Index.
- This raised discussions about the importance of following best practices in AI safety amidst concerns about potential model risks.
- Tech Leaders urged to prioritize country interests: In a public exchange, a tech leader encouraged Elon Musk to consider national interests, mentioning the development of new sites and acknowledging potential conflicts with corporate goals.
- This sentiment highlights the ongoing conversation about the responsibilities of influential figures in the tech industry to prioritize societal benefits.
Links mentioned:
- Tweet from NIK (@ns123abc): 🚨🚨🚨BREAKING: Microsoft CEO was just asked about @elonmusk saying Project Stargate doesn't have the money to invest“All I know is, I’m good for my $80 Billion”LMFAOOO
- Tweet from Sam Altman (@sama): @elonmusk @OpenAI wrong, as you surely know.want to come visit the first site already under way?this is great for the country. i realize what is great for the country isn't always what's optim...
- openbmb/MiniCPM-o-2_6 · Safety: no description found
Interconnects (Nathan Lambert) ▷ #random (88 messages🔥🔥):
Robonato humor, OpenAI Media Manager developments, DeepSeek advancements, Creative writing benchmarks, AI podcast editing tools
-
Robonato humor takes off: Members discussed humorous takes on Robonato, with one user sharing a hilarious video featuring Sam Altman and Alexandr Wang, noting the amusing moment when a chair broke during a major event.
- The conversation highlighted the community's craving for more content around Robonato.
- OpenAI Media Manager is on the horizon: A discussion revealed that the OpenAI Media Manager tool is still under development, with its release timing remaining uncertain.
- Listeners were updated on a recent interview with OpenAI's Chief Product Officer, who mentioned advancements like o3-mini and anticipated AI agents.
- DeepSeek's impressive capabilities: Users debated the advancements of DeepSeek, particularly the DeepSeek-R1 model's success in creative writing and humor, which seems to outperform other models.
- Discussions also emphasized the historical insights gained from DeepSeek's outputs compared to other models, particularly when analyzing complex topics.
- Emotional Intelligence Benchmarks for LLMs: The launch of EQ-Bench, focusing on emotional intelligence benchmarks for LLMs and creative writing, attracted interest from members looking into its potential implications.
- Users shared enthusiasm for the benchmarks while discussing the performance of various models in creative contexts.
- Podcast editing tools and insights: A member expressed interest in exploring AI-powered podcast editing tools, debating their effectiveness and the potential return on investment that could be achieved.
- The topic brought up the challenges of producing quality audio-visual content while balancing time demands.
Links mentioned:
- EQ-Bench Creative Writing Leaderboard: no description found
- Tweet from Eric Hartford (@cognitivecompai): I got a sponsor! Thanks @driaforall! The data will be published to @huggingface with Apache-2.0 license, in a couple days.
- Tweet from thebes (@voooooogel): no description found
- Tweet from Ashlee Vance (@ashleevance): Watcha guys think? $2,500 to have one of my books sucked into an LLM. Yes, no?
- Tweet from Eric Hartford (@cognitivecompai): $6k in API fees to create Dolphin-R1 dataset. I follow the Deepseek-R1 distillation recipe, but with Dolphin seed data. (600k of reasoning, 200k of chat, 800k total)I want to license it Apache 2.0, ...
- Tweet from thebes (@voooooogel): no description found
- Tweet from nano (@nanulled): the longer r1 thinks the more alien it becomes.you start to notice some weird token preferencesit prefers to save some space and write numbers and letters togetherwe could actually see a more efficien...
- Tweet from Tibor Blaho (@btibor91): OpenAI Chief Product Officer Kevin Weil interview at Journal House in Davos, Switzerland- OpenAI will launch o3-mini "very soon" followed by full o3 in "February, March, if everything goes...
- The AI Rights licensing platform for creators - Created by Humans: Take control of your works' AI Rights and get compensated for their use by AI companies.
- Tweet from Nathan Lambert (@natolambert): Change my mind: The safety work on o1 from OpenAI is already a ton more evidence than people are giving credit for that reasoning-heavy training will generalize benefits to other domains.Safety is sup...
- Tweet from spor (@sporadicalia): i am absolutely losing my mind at Sam Altman and Alexandr Wang in the background of this videoQuoting DramaAlert (@DramaAlert) Theo Von's chair BROKE midway through Trump's Inauguration and fe...
- Tweet from Dean W. Ball (@deanwball): ok so this one is giving away a little bit about a forthcoming piece of mine, but also niche and purely non-technical--in fact, it's nearly a pure humanities question. the prompt was: "did bee...
- Tweet from thebes (@voooooogel): r1 can draw spirals!that may not sound like a big deal, but other models (including o1) struggle with this quite a bit for some reason. r1 successfully draws a spiral roughly half the time.
- The Greatest Shot In Television: This is the single greatest shot filmed in television history. It's completely real, it's NOT a green-screen. One chance, and if James Burke had missed, ther...
- "happy claude, free from trolley problems" Sticker for Sale by vgel: Buy "happy claude, free from trolley problems" by vgel as a Sticker
- - YouTube: no description found
- DeepSeek创始人梁文锋,广东人,仅靠百名中国程序员,赶超OpenAI_腾讯新闻: 今天介绍一位金融和人工智能领域的创业者梁文锋,他是幻方和深度求索(DeepSeek)两家公司的创始人。 即刻网友@Chris-Su对梁文锋的评价我觉得很到位: “梁文锋是极少数还....
Interconnects (Nathan Lambert) ▷ #memes (4 messages):
AI Safety Index, Whitehouse.com betting concerns, Undefined discussions, Contrast in social media posts
-
Debate Over AI Safety Index: A user referenced a conversation about measuring the safety of AI models using the AI Safety Index, highlighting a casual remark questioning AI's potential dangers.
- What are you afraid about? That it starts eating the user remotely? was a humorous community response during this exchange.
- Concerns about dubious emails and websites: A member expressed concerns over an email linked to whitehouse.com, labeling it a potentially illegal betting website based on its appearance.
- This was accompanied by an image screenshot raising questions among the community regarding its legitimacy.
- Social Media Contrast Narratives: A member pointed out the amusing contrast in a social media post without providing further context, leading to a light-hearted commentary.
- Comments around this post included reactions on the peculiarities of social media content and its impact.
- Undefined Content Frustration: A user stumbled upon a link that contained the term 'undefined', provoking confusion and prompting a demand for clarity.
- This sparked brief dialogue around what could be behind undefined discussions in tech circles.
Links mentioned:
- Tweet from undefined: no description found
- Tweet from Tanishq Mathew Abraham, Ph.D. (@iScienceLuvr): the contrast lol
- Tweet from Yifei Hu (@hu_yifei): > "Is your model safe? You could try measuring it using this AI Safety Index (and linked in the comments)> some random dude: what are you afraid about ? that it starts eating the user remote...
Interconnects (Nathan Lambert) ▷ #rl (13 messages🔥):
GRPO concerns, T1 RL paper, Deepseek response, HLF training plots, TRL GitHub discussion
-
Concerns about GRPO implementation: A member expressed skepticism about the GRPO implementation, stating it lacks PPO clipping and questioning the application of KL divergence directly to loss rather than rewards.
- They mentioned, 'Seems like the KL is supposed to be applied to loss in GRPO' and highlighted the potential implications of this approach.
- Introduction of T1 RL Scaling Method: A new paper from Ziphu and Tsinghua introduces T1, a method aimed at scaling reinforcement learning for large language models with a focus on enhancing sampling diversity.
- The paper includes a detailed overview of how trial-and-error and self-verification are synthesized in RL training, as discussed in the arXiv paper.
- Deepseek's Enthusiasm: In response to a query about the Deepseek paper, a member humorously indicated it was well-received, suggesting that 'deepseek likes it 🤣'.
- This lighthearted comment indicates a positive reception within the community towards the paper's ideas.
- Popularity of RL Training Plots: Discussion arose over the increasing popularity of RL training-time scaling plots, with a member noting that it isn't just a comparison against RL steps.
- They acknowledged the plot as 'cool,' signaling interest in the visual representation of RL progress metrics.
- Discussion on GRPO within Hugging Face's TRL: A GitHub issue was shared regarding questions on the GRPO implementation in Hugging Face's TRL, raising concerns and inviting collaboration.
- The member mentioned gathering insights on applying KL distance in advantages, referencing the relevant issue for further discussion.
Links mentioned:
- Advancing Language Model Reasoning through Reinforcement Learning and Inference Scaling: Large language models (LLMs) have demonstrated remarkable capabilities in complex reasoning tasks. However, existing approaches mainly rely on imitation learning and struggle to achieve effective test...
- GRPO questions · Issue #2608 · huggingface/trl: Hey friends! I have some questions on the GRPO implementation, happy to discuss. It looks like you apply the KL distance in the advantages, while the DeepSeekMath paper says “Also note that, instea...
Interconnects (Nathan Lambert) ▷ #reads (12 messages🔥):
Davos Interviews, AI Regulations, Transistor Radio Podcasts
-
Dario Amodei discusses Claude at Davos: In a YouTube interview, Anthropic CEO Dario Amodei outlines the future of Claude, including features like web browsing and voice capabilities.
- Poor Dario in Davos while Sama is in the White House with Donny highlighted the contrasting settings of tech leaders.
- New AI Regulations explored: In another YouTube session, Dylan Patel from SemiAnalysis breaks down the implications of recent AI regulations, discussing the winners and losers in this evolving landscape.
- This episode, while recorded on January 15, focuses on the current diffusion rules and country tier list.
- Puffy Vests Sighted: A humorous observation noted that the speaker typically watches events to see what puffy vests notable figures like Alex Karp are wearing.
- This illustrates the lighter side of following tech conferences, even amid serious discussions.
- AI Development Tools Shared: A member shared a link to GitHub resources where developers can build AI apps and agents leveraging OpenAI, Anthropic, and others.
- They also suggested trying it out on Hugging Face here.
- Recommendation for Unhinged Podcasts: A recommendation was made for Unhinged podcasts, specifically Transistor Radio, indicating a preference for more offbeat content.
- One user humorously noted their need for additional unorthodox listening options.
Links mentioned:
- Tweet from AK (@_akhaliq): @OpenAI awesome, while we wait, developers can build ai apps and agents with openai, anthropic, google, nvidia and more here: https://github.com/AK391/ai-gradiousers can try it out here: https://huggi...
- - YouTube: no description found
- Inside Anthropic's Race to Build a Smarter Claude and Human-Level AI | WSJ: At WSJ Journal House Davos, Anthropic CEO Dario Amodei outlines Claude’s next chapter—from web browsing, voice to more advanced models—while predicting that ...
- New AI Regulations Winners & Losers with SemiAnalysis’s Dylan Patel: In this episode of Unsupervised Learning, we sit down with Dylan Patel, Chief Analyst at SemiAnalysis, to break down what these sweeping changes really mean....
Interconnects (Nathan Lambert) ▷ #posts (3 messages):
SnailBot News, Bot Performance
-
SnailBot News Notification: A notification was sent out for SnailBot News, tagged to a specific role in the Discord server.
- The message indicates that the Bot is ready to share updates with members.
- Discussion on SnailBot Speed: A member noted that SnailBot's responsiveness was not too slow, suggesting improved performance.
- This comment reflects a positive perception of the Bot's speed in delivering updates.
Interconnects (Nathan Lambert) ▷ #policy (4 messages):
Arms Race in Technology, Live Events, Stargate Discussions
- Ongoing Arms Race in AI: My expectation is that this is an arms race and there is no reachable state of the world where it isn't. The discussions imply a consensus that the nature of competition in AI is relentless and ever-evolving.
-
Live Broadcast Alert: A member shared a link to a live stream, indicating that they were currently broadcasting.
- This sparked interest among others, leading to inquiries about whether anyone could catch it live.
- Missed Opportunity with Stargate: There was a lament about missing a live event related to Stargate, with the member expressing that it would have been electric to catch live.
- This showcases a shared enthusiasm for significant media events within the community.
aider (Paul Gauthier) ▷ #general (122 messages🔥🔥):
Gemini 2.0 Flash Thinking Model, Aider Workflow Enhancements, Model Comparison and Critique, Markdown Specifications for Development, RAG Approach with PDF References
-
Discussion on Gemini 2.0 Flash Thinking Model: A user reported that Google released the
gemini-2.0-flash-thinking-exp-01-21
with a 1 million token context window and 64K max output tokens.- Users are curious about its performance compared to established models like DeepSeek R1, with some engaging in tests.
- Iterative Workflow Enhancements in Aider: A user shared a workflow involving the use of markdown documents for major features and refactoring, enhancing collaboration with LLMs.
- This approach emphasizes self-assessment from LLMs and creating comprehensive specifications to improve code quality and speed up development.
- Evaluating Model Performance and Critiques: Members discussed the differences between various models, noting that Sonnet often proposes better solutions compared to R1.
- Users suggest integrating reasoning outputs from R1 into other models for improved performance, particularly with complex scenarios.
- Using Markdown for Specification and Unit Tests: Another user advocates for writing specifications in markdown for generating unit tests, resulting in better code quality and project management.
- This technique allows for more efficient development processes, ensuring alignment with specified requirements throughout the coding cycle.
- Adding Context and Resources in Aider: A user inquired about referencing PDFs in Aider for a RAG approach, which is supported with Sonnet through a simple command.
- This highlights the growing interest in optimizing Aider's functionality to utilize external resources effectively.
Links mentioned:
- /2025/01/memory-makes-computation-universal/: no description found
- Secure Llm Leaderboard - a Hugging Face Space by stacklok: no description found
- Using /help: Use “/help " to ask for help about using aider, customizing settings, troubleshooting, using LLMs, etc.
- Tweet from Logan Kilpatrick (@OfficialLoganK): We are rolling out a new Gemini 2.0 Flash Thinking update:- Exp-01-21 variant in AI Studio and API for free- 1 million token context window- Native code execution support- Longer output token generati...
- Inference Catalog | Inference Endpoints by Hugging Face: no description found
- Reasoning Model (deepseek-reasoner) | DeepSeek API Docs: deepseek-reasoner is a reasoning model developed by DeepSeek. Before delivering the final answer, the model first generates a Chain of Thought (CoT) to enhance the accuracy of its responses. Our API p...
- no title found: no description found
aider (Paul Gauthier) ▷ #questions-and-tips (91 messages🔥🔥):
Aider model configurations, Error handling in Aider, Using OpenAI keys with Aider, Integrating various models, Neovim plugins for Aider
-
Issues with Aider configuration: Users reported difficulties configuring Aider, particularly with model settings such as
r1: true
not functioning as expected.- It was suggested to use
--model r1
instead and to check usingaider --verbose
for more diagnostic information. - Encountering upgrade errors in Aider: Several users encountered persistent upgrade prompts or errors that prevent Aider from functioning properly after various installation attempts.
- Recommendations included checking and potentially deleting the
.aider
directory and re-installing via simpler methods likepip install
. - Using LLMs with Aider: Discussion emerged around the necessity of LLM API keys to utilize Aider's editing capabilities, prompting questions about flexibility in using various web-based chat services.
- It was noted that copy and paste mode requires a defined LLM for operation, limiting usage to those with API access.
- Error reporting connections to deep learning models: Individuals shared their experiences of getting 0% pass rates when benchmarking various versions of models with Aider, indicating troubleshooting requirements.
- Suggestions to investigate quantized weights and detailed logging of model installations were provided to address potential issues.
- Neovim support for Aider: A query was raised regarding the best Neovim plugin for integrating Aider functionalities within the editor's environment.
- Details and recommendations for effective Neovim setups for Aider users remain an area of evolving discussion.
- It was suggested to use
Links mentioned:
- Providers | liteLLM: Learn how to deploy + call models from different providers on LiteLLM
- Installation: How to install and get started pair programming with aider.
- Release history: Release notes and stats on aider writing its own code.
- Advanced model settings: Configuring advanced settings for LLMs.
- Advanced model settings: Configuring advanced settings for LLMs.
- [Feature]: DeepSeek-R1 support · Issue #7877 · BerriAI/litellm: The Feature DeepSeek-R1 API returns its thoughts inside the reasoning_content parameter. Currently this is ignored by LiteLLM. Their API approach, of return "reasoning_content" for the long-...
- Aider bench 1.5B (AWQ) Python subset, diff format: Aider bench 1.5B (AWQ) Python subset, diff format. GitHub Gist: instantly share code, notes, and snippets.
- r1-cli/VLLM.md at master · lmmx/r1-cli: Simple CLI to use DeepSeek r1 on the command line (32B in 4-bit via Unsloth on Transformers) - lmmx/r1-cli
- r1-cli/TGI.md at master · lmmx/r1-cli: Simple CLI to use DeepSeek r1 on the command line (32B in 4-bit via Unsloth on Transformers) - lmmx/r1-cli
aider (Paul Gauthier) ▷ #links (1 messages):
astor1: https://github.com/NixOS/nixpkgs/pull/375634 aider 0.72.1 in nixpkgs
Stackblitz (Bolt.new) ▷ #announcements (1 messages):
Bolt funding, Community appreciation
-
Bolt secures $105.5M in Series B Funding: Today, Bolt announced $105.5 million in funding to enhance its capabilities, led by Emergence & GV, with contributions from investors including Madrona and The Chainsmokers.
- This significant investment aims to propel Bolt to new heights in the devtools and AI sectors.
- Gratitude for the Community: The team expressed deep appreciation for the amazing community, stating it wouldn't be possible to move forward without them.
- They emphasized their commitment to making Bolt even more powerful for users as they progress.
Link mentioned: Tweet from bolt.new (@boltdotnew): Today we're announcing $105.5m in funding to take Bolt to new heights! 🚀Our Series B was led by Emergence & GV, with participation from Madrona, The Chainsmokers (Mantis), Conviction, and some of...
Stackblitz (Bolt.new) ▷ #prompting (16 messages🔥):
Netlify routing issues, NextJS and Supabase integration, SSR challenges with large NextJS projects, Building a Tetris mini app for Telegram
-
Netlify's Routing Woes Discovered: Members discussed difficulty with Netlify routing, specifically encountering a 404 page when trying to access routes directly like /Imprint.
- One suggestion was made to create a _redirects file to solve the page routing issues.
- NextJS Sparks Debates: A developer shared that their stack of NextJS, shadcn, and Tailwind works well for UX despite persistent integration issues with Supabase.
- Concerns were raised about NextJS not functioning with webcontainers, leading to a call for guidance on its current viability.
- SSR Struggles with NextJS Deployments: A member is combating SSRs challenges while deploying their large NextJS project, overwhelmed by importing hundreds of blog posts during build time.
- Concerns were expressed over Netlify's 10 second timeout on the free plan and doubts about upgrading to the pro plan for additional build time.
- Seeking Solutions with Prismic and Contentlayer: While assessing options for content sources, a member explored using Prismic, but realized it isn’t suitable for container-based environments like Bolt.
- They are now investigating Contentlayer as an alternative and shared a YouTube video for insights.
- Creating a Unique Tetris Telegram Mini App: A member is seeking help to develop a Telegram mini app for a Tetris game, but wants to replace bricks with slices of meat.
- They aim to implement a leaderboard and referenced the Telegram Apps Center as a potential platform for their application.
Link mentioned: Telegram Apps Center: Community-driven catalog of applications developed by third-party developers. Not affiliated with Telegram Messenger.
Stackblitz (Bolt.new) ▷ #discussions (192 messages🔥🔥):
Bolt recursion issues, Token upgrade inquiries, User permissions and policies, CORS issues, Using Claude for troubleshooting
- Bolt's recursion infinite loop bug: Users are experiencing infinite loops when dealing with user policies in Supabase, which Bolt struggles to resolve. A suggested workaround involves reviewing trigger functions created by Bolt.
- Questions about upgrading subscription plans: Users inquired about upgrading their subscription plans, particularly how tokens would be allocated upon upgrade. It’s confirmed that current users will receive a pro-rated amount of tokens to finish the current cycle.
- Challenges with user account management: Several users expressed difficulties with managing user account types and permissions in Supabase through Bolt, highlighting significant token usage in troubleshooting. The process involves defining clear policies, but issues persist due to Bolt’s inability to handle complexities.
- CORS issues with Supabase edge functions: Users are encountering issues with CORS when using Supabase edge functions, leading to requests being blocked. Recommendations suggest ensuring correct request configurations and possibly utilizing Node for API requests.
- Utilizing Claude for problem-solving: Users found success in using Claude to tackle specific coding problems and retrieve database policy updates. This collaborative approach is seen as essential when Bolt fails to address coding issues directly.
Links mentioned:
- Vettivel Bank (BETA) - StackBlitz: A personal finance tracker.
- Loom | Free Screen & Video Recording Software | Loom - 22 January 2025: Use Loom to record quick videos of your screen and cam. Explain anything clearly and easily – and skip the meeting. An essential tool for hybrid workplaces.
- Postman: The World's Leading API Platform | Sign Up for Free: Accelerate API development with Postman's all-in-one platform. Streamline collaboration and simplify the API lifecycle for faster, better results. Learn more.
- READ THIS FIRST: Critical information about Bolt.new's capabilities, limitations, and best practices for success
- Cursor Directory: Find the best cursor rules for your framework and language
Yannick Kilcher ▷ #general (161 messages🔥🔥):
R1's Performance, DeepSeek's Advancements, AI Ethical Dilemmas, OpenAI vs. Competitors, AI Infrastructure Investments
-
R1 impresses with performance and challenges: Users reported mixed results from the R1 model, with some impressed by its ability to tackle complex questions while others highlight its struggles with simpler tasks.
- Discussion around contextual tokenization and potential issues in the model's reasoning capabilities was noted.
- DeepSeek seen as a competitor to OpenAI: DeepSeek is praised for its high-quality outputs at a much lower cost compared to OpenAI's offerings, leading to speculation about its potential to disrupt the AI model market.
- Users emphasize the importance of features and control, implying that DeepSeek might lure businesses away from OpenAI.
- Ethical dilemmas of AI and job displacement: A user expressed moral concerns over the potential job losses caused by the success of their AI startup, questioning whether this makes them evil.
- Another user likened this dilemma to the everyday conveniences that lead to job losses, suggesting a broader philosophical discussion.
- The rivalry between OpenAI and Elon Musk: Tensions rise with accusations regarding allegiance to America and the implications of OpenAI's partnerships with government firms.
- Some users suggested that traditional companies often default to closed-source solutions despite the potential benefits of open-source alternatives.
- Enormous investments in AI infrastructure: The Stargate Project, announced to invest $500 billion into AI infrastructure, sparked discussions regarding its implications for American technology leadership.
- Concerns were raised over corporate influences in government and the prioritization of profit over broader social implications.
Links mentioned:
- Tweet from AshutoshShrivastava (@ai_for_success): 🚨 $500 BILLION FOR AGI !!!! Biggest AI Project ever, Trump announces $500 billion 'Stargate' AI project for AI and re-industrialization of the United States. - SoftBank oversees finances, Op...
- Tweet from Tsarathustra (@tsarnick): President Trump announces the Stargate Project, the largest AI infrastructure investment in history at $500 billion, to build "colossal data centers"
- Tweet from AshutoshShrivastava (@ai_for_success): Elon vs. Sam Altman is getting nasty. Is Sam Altman trying to imply that Elon and his companies don’t put America first, while OpenAI does?Let’s not forget that OpenAI’s board now includes BlackRock e...
- Tweet from Chubby♨️ (@kimmonismus): https://x.com/ns123abc/status/1881986668238168563/video/1 OpenAI Investor Masayoshi Son: "I think AGI is coming very, very soon. And after that aritificial superintelligence to solve the issues m...
- Tweet from Paul Calcraft (@paul_cal): OpenAI's @SebastienBubeck on the o1 paradigm:"No tactic was given to the model. Everything is emergent. Everything is learned through reinforcement learning. This is insane. Insanity"Quoti...
- Tweet from Beff – e/acc (@BasedBeffJezos): We are in the Golden Age of techno-capital acceleration.Strap in, folks. 🚀🙌🔥🇺🇸
- DeepSeek R1 - API, Providers, Stats: DeepSeek-R1 is here!⚡ Performance on par with OpenAI-o1📖 Fully open-source model & technical report🏆 MIT licensed: Distill & commercialize freely!. Run DeepSeek R1 with API
- Tweet from OpenAI (@OpenAI): Announcing The Stargate ProjectThe Stargate Project is a new company which intends to invest $500 billion over the next four years building new AI infrastructure for OpenAI in the United States. We wi...
- GitHub - microsoft/aici: AICI: Prompts as (Wasm) Programs: AICI: Prompts as (Wasm) Programs. Contribute to microsoft/aici development by creating an account on GitHub.
- AI Website Generator » SiteForge: no description found
Yannick Kilcher ▷ #paper-discussion (32 messages🔥):
DeepSeek R1 Model Performance, Challenges in Paper Reviewing, DeepSeekMath Paper Insights, Anthropomorphism in Models, DeepSeek Training Procedure
-
DeepSeek R1 outshines other models: Members discussed how DeepSeek R1 performed at 92% while other models like Gemini and O1 scored significantly lower in math-related evaluations.
- Bojan highlighted that R1 excels due to its geometric understanding and analogies drawn from various models.
- Recruiting Reviewers proves challenging: A member expressed frustration about needing to personally recruit 12 reviewers out of 50+ interested parties to ensure top-quality reviews.
- This raised concerns about the review process in top conferences and the effort required to maintain quality standards.
- DeepSeekMath and GRPO: The community is set to discuss the paper DeepSeekMath and its mathematical reasoning techniques using GRPO in a future event.
- Members noted the paper's length of 22 pages, but expressed optimism about navigating through its contents.
- Critique on Anthropomorphism in AI: Some members criticized the paper's reliance on anthropomorphism, suggesting that it makes the model's reasoning appear more human than it is.
- Bojan pointed out that training on human data leads to models mimicking human-like reasoning patterns.
- DeepSeek Training Process Unlocked: A detailed breakdown of the DeepSeek R1 training procedure was provided, showcasing the multi-stage training loop involving RL and finetuning.
- The discussion highlighted the potential performance improvements through structured stages of training, with valuable insights shared by the community.
Links mentioned:
- DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models: Mathematical reasoning poses a significant challenge for language models due to its complex and structured nature. In this paper, we introduce DeepSeekMath 7B, which continues pre-training DeepSeek-Co...
- Thread by @casper_hansen_ on Thread Reader App: @casper_hansen_: The DeepSeek R1 training procedure confused me at first. My brain refused to accept this powerful model could be incredibly straightforward. Let me break down this elegant beast for y...
- DeepSeek-R1/DeepSeek_R1.pdf at main · deepseek-ai/DeepSeek-R1: Contribute to deepseek-ai/DeepSeek-R1 development by creating an account on GitHub.
Yannick Kilcher ▷ #agents (1 messages):
IntellAgent, Conversational Agents Evaluation, Synthetic Interactions
-
Introducing IntellAgent Framework: The new open-source project, IntellAgent, is a framework designed for comprehensive diagnosis and evaluation of conversational agents using simulated, realistic synthetic interactions. You can check it out on GitHub.
- This cutting-edge framework generates diverse datasets from the agent's prompt, simulating conversations, and providing detailed critiques.
- New Insights from Research Paper: The accompanying research paper reveals several fascinating and non-trivial insights produced by the IntellAgent system, available for reading here.
- Insights are derived from the performance and evaluation of conversational agents using this innovative framework.
- Visual Overview of IntellAgent System: A visual representation titled intellagent_system_overview.gif illustrates the workings of the IntellAgent framework.
- This overview provides an engaging insight into the components and functionality of the evaluation process.
Link mentioned: GitHub - plurai-ai/intellagent: A framework for comprehensive diagnosis and evaluation of conversational agents using simulated, realistic synthetic interactions: A framework for comprehensive diagnosis and evaluation of conversational agents using simulated, realistic synthetic interactions - plurai-ai/intellagent
Yannick Kilcher ▷ #ml-news (3 messages):
Stargate Project, UI-TARS Model, OpenAI Operator Feature
- OpenAI Announces Stargate Project: OpenAI revealed details about the Stargate Project which focuses on enhanced AI interactions. More information can be found in their announcement.
-
Hugging Face's UI-TARS Model Released: The UI-TARS model, designed for automated GUI interactions, is now available on Hugging Face. This follows the release of the paper on UI-TARS: Pioneering Automated GUI Interaction with Native Agents.
- The repository includes various versions such as UI-TARS-2B-SFT and UI-TARS-7B-DPO.
- OpenAI's Upcoming Operator Feature: Scoop reveals that OpenAI is preparing to release a new feature called Operator for ChatGPT, which will take actions on behalf of users in their browsers. It will include suggested prompts and the ability to save/share tasks, but it won't be available in the API, as reported here.
Links mentioned:
- Tweet from Stephanie Palazzolo (@steph_palazzolo): Scoop: OpenAI is prepping to release "Operator," a new ChatGPT feature that will take actions on behalf of users in their browsers, this week.Interesting details:- Operator provides suggested ...
- bytedance-research/UI-TARS-7B-SFT · Hugging Face: no description found
- UI-TARS - a Hugging Face Space by Aheader: no description found
OpenRouter (Alex Atallah) ▷ #announcements (1 messages):
Web Search Pricing, API Access Launch
-
Web Search Pricing Set at $4/1k Results: A new pricing model for web search has been established at $4/1k results, set to be implemented starting tomorrow.
- Each request will typically include up to 5 web search results, resulting in an approximate cost of less than $0.02 per request.
- API Access to Soft Launch Tomorrow: The API access is set to soft launch alongside the new pricing, introducing expanded customizability options.
- Members are encouraged to share any feedback or questions regarding the new features.
OpenRouter (Alex Atallah) ▷ #general (187 messages🔥🔥):
DeepSeek model performance, DeepSeek R1 issues, Censorship concerns, Uncensored models, Cerebras model availability
-
DeepSeek R1 experiences performance drop: Multiple users reported an 85% drop in API performance for DeepSeek R1 overnight, hinting at increased scrutiny on this provider.
- Concerns were raised about potential censorship within the model, particularly when discussing sensitive topics.
- Concerns over censorship in AI models: Users expressed frustration with the censorship found in models like DeepSeek R1 and discussed how to navigate it.
- Tricks to bypass censorship were shared, with humor about the risks involved in pressing boundaries on uncensored content.
- Favorite uncensored models on OpenRouter: Discussion centered around which models on OpenRouter are considered uncensored and coherent, with users mentioning options like Dolphin models and Hermes.
- For NSFW roleplaying, choices were narrowed down further, indicating limited yet popular selections.
- Availability of Cerebras models: Despite mentions of Cerebras' Mistral Large, users confirmed it is not generally available for public use, leading to frustration about inaccessible models.
- Many noted that only Llama models seem to be available from Cerebras, raising questions about the validity of model existence claims.
- Upcoming feature fixes and enhancements: OpenRouter team confirmed they are working on resolving issues with DeepSeek R1 and the upcoming API for R1 search.
- Users are encouraged to keep an eye out for improvements, with ongoing transparency regarding maintenance and updates.
Links mentioned:
- Hyperbolic AI Dashboard: no description found
- DeepSeek R1 – Uptime and Availability: Uptime statistics for DeepSeek R1 across providers - DeepSeek-R1 is here!⚡ Performance on par with OpenAI-o1📖 Fully open-source model & technical report🏆 MIT licensed: Distill & commercializ...
- Hyperbolic | OpenRouter: Browse models provided by Hyperbolic
- OpenRouter: A unified interface for LLMs. Find the best models & prices for your prompts
- no title found: no description found
- no title found: no description found
- Fireworks - Fastest Inference for Generative AI: Use state-of-the-art, open-source LLMs and image models at blazing fast speed, or fine-tune and deploy your own at no additional cost with Fireworks AI!
- Hyperbolic AI Dashboard: no description found
- 账号登录-火山引擎: no description found
- no title found: no description found
Perplexity AI ▷ #announcements (1 messages):
Sonar API, Sonar Pro, AI Companion 2.0, SimpleQA benchmark, Data security
-
Launch of Sonar and Sonar Pro API: Today, Perplexity introduced the new Sonar and Sonar Pro API, empowering developers to create apps with generative search capabilities backed by real-time web research and robust citation features.
- Companies utilizing Sonar have reported significant successes, with Zoom incorporating the API to enhance its AI Companion 2.0 product.
- Sonar Pro beats major competitors: Recent findings from the SimpleQA benchmark revealed that Sonar Pro outperformed leading search engines and LLMs in terms of answer quality.
- Web-wide research and Q&A capabilities provided by Sonar are noted as being unparalleled.
- Affordability of Sonar’s API: The pricing for Sonar's grounding requests is claimed to be more affordable than competing offerings, allowing users to power products with the fastest, cheapest API.
- Perplexity asserts that you can start building and integrating their tech within just a few minutes.
- Data privacy assurance from Perplexity: A significant highlight is that Perplexity does not conduct LLM training on user data, ensuring privacy and security.
- This aspect is crucial for developers concerned about the safety of their data when using machine learning APIs.
Link mentioned: Sonar by Perplexity: Build with the best AI answer engine API, created by Perplexity. Power your products with the fastest, cheapest offering out there with search grounding. Delivering unparalleled real-time, web-wide re...
Perplexity AI ▷ #general (128 messages🔥🔥):
Sonar API Performance Issues, Pro Model Usage Confusion, Model Comparison and Updates, Login and Server Errors, API Functionality and Documentation
-
Sonar API performance struggles highlighted: Users reported various errors with the Sonar Pro API including 500, 402, and 403 errors while trying to integrate it into their projects, expressing frustration with its functionality.
- Some users speculated that the API could be down or experiencing capacity issues, leading to a temporary halt in performance.
- Confusion over Pro model access: Multiple users inquired about the steps to activate the Pro model via the API and whether certain roles or capabilities apply to their subscriptions.
- Discussions pointed out the complexity of properly using the Pro features, including necessary modifications to calls for desired model outputs.
- Comparisons between Sonar models: Feedback indicated that Sonar Large is now outperforming Sonar Huge, prompting users to seek updated information on the model line.
- One user expressed disappointment over the impending removal of Sonar Huge from the available options, illustrating the concerns about speed versus performance.
- Login issues and server capacity: A number of users reported experiencing 500 Internal Server Errors and other login difficulties while trying to access Perplexity on both mobile and desktop.
- Some believed that these issues might be linked to server capacity constraints, as many users were unable to connect.
- API usage guidance requested: Several users sought assistance with navigating the API documentation for Perplexity Pro, asking for clarity on the parameters and functions available.
- Helpful links to documentation were shared among users to facilitate better understanding and usage of the API features.
Links mentioned:
- Supported Models - Perplexity: no description found
- Tweet from Henry Modisett (@henrymodis)): Very excited about this launch but i'm particularly excited for this new sub-brand developed by one of our designers (Erin McKnight). She has expanded our universe!Quoting Perplexity (@perplexity_...
- GitHub - PierrunoYT/perplexity-webui: A modern web interface for interacting with the Perplexity AI API.: A modern web interface for interacting with the Perplexity AI API. - PierrunoYT/perplexity-webui
Perplexity AI ▷ #sharing (8 messages🔥):
Perplexity API usage, Anduril's autonomous weapons, PhD-level Super Agents, College basketball, pyctc_decode for projects
-
Perplexity API Query: Several users inquired about the possibility of using the Perplexity API for their projects, referencing a specific search topic here.
- A high level of interest was observed concerning practical applications of the API for various tasks.
- Anduril's $1B Weapons Factory: Discussion highlighted Anduril's new $1B Autonomous Weapons Factory, emphasizing its potential impact on defense technologies in a video shared here.
- The topic raised questions among users about autonomous warfare and ethical implications.
- Briefing on ‘PhD-level Super Agents’: There was mention of Altman briefing D.C. officials on emerging technologies like ‘PhD-level Super Agents’, hinting at advanced AI capabilities.
- Members speculated on what these capabilities might entail for future AI applications.
- Search Inquiry on College Basketball: A user sought information regarding college basketball by referencing a specific search link, indicating an inquiry into statistical data.
- The relevance of sports statistics in AI applications was briefly discussed.
- Exploring pyctc_decode: A user shared their ongoing exploration of pyctc_decode for a project, reflecting the practical application of this tool.
- This generated interest in how pyctc_decode can enhance project outcomes.
Link mentioned: YouTube: no description found
Perplexity AI ▷ #pplx-api (16 messages🔥):
Sonar Pro API, Search Domain Filter, Error Messages, Deployment in Europe, Comparison Tool
-
Search Domain Filter Functionality Under Scrutiny: A user questioned if the search_domain_filter works with the new Sonar Pro as it seems to be ignored without error message, which raised concerns about its reliability.
- Another member noted that the Search Domain Filter is a tier 3 beta feature, implying potential stability issues.
- Community Responds with Error Message Updates: A user suggested that the team should add a more helpful error message when the search_domain_filter is not working, which was acknowledged by community members.
- The team is actively working on this, showing responsiveness to user feedback regarding Sonar API functionality.
- GDPR Compliance Concerns for European Deployment: A member inquired about the deployment of Sonar Pro in Europe to ensure GDPR compliance, expressing urgency for local server integration.
- This highlights the growing demand for compliant AI solutions amidst increasing regulatory scrutiny.
- API Performance Issues Post-Sonar Launch: Several users reported experiencing 524 errors when attempting to switch to the Sonar API, indicating potential reliability issues following its launch.
- Community members confirmed similar experiences, suggesting a temporary spike in usage causing these latency problems.
- GitHub Tool for Model Comparison Emerges: A user shared a link to a GitHub tool designed to compare different Perplexity AI models, enhancing usability for developers.
- The tool allows side-by-side comparisons of models, contributing to a better understanding of their functionalities and performance metrics.
Links mentioned:
- no title found: no description found
- Rate Limits and Usage Tiers - Perplexity: no description found
- Tweet from Henry Modisett (@henrymodis)): Very excited about this launch but i'm particularly excited for this new sub-brand developed by one of our designers (Erin McKnight). She has expanded our universe!Quoting Perplexity (@perplexity_...
- GitHub - jsandai/pplx-api-compare: A modern React-based web application for comparing different Perplexity AI models side by side. This tool allows you to test prompts across multiple models simultaneously and compare their responses, token usage, and costs.: A modern React-based web application for comparing different Perplexity AI models side by side. This tool allows you to test prompts across multiple models simultaneously and compare their response...
MCP (Glama) ▷ #general (114 messages🔥🔥):
MCP Server Functionality, Brave Search for Documentation, Custom GPT Limitations, Code Editing with MCP, Prompt System in Claude Desktop
-
MCP Server Functionality Has Room for Improvement: Users are discussing the limitations of custom GPTs and the ease of using MCP servers for various tasks, noting that MCP's editing capabilities for complex codebases could improve.
- One member suggested that having different LLMs for different purposes might offer advantages, while another pointed out that mixing tools with semantic selection could streamline processes.
- Exploring Brave Search for Latest Documentation: Members are utilizing Brave Search to fetch documentation and dependencies, with one noting the ability to scrape docs and compile them into markdown files.
- While experimenting with the brave-search MCP server, users shared snippets and methods for effective implementation, highlighting the potential for automation.
- Challenges with Custom GPTs in 2024: Concerns were raised regarding the lack of improvement for custom GPTs, as they fail to integrate newer ChatGPT features effectively.
- Discussions included the potential devaluation of these models and the disappointment over unmet expectations surrounding the custom GPT marketplace.
- Code Editing Features of MCP: Members are interested in the code editing feature of MCP servers, discussing its limitations in handling large codebases and potential for improved functionality.
- There are thoughts towards converting MCP servers into functions for better interaction, with emphasis on utilizing git versioning for tracking changes.
- Implementing Custom Prompts in Claude Desktop: Users are struggling with how to make prompts available via the prompts/List endpoint in Claude Desktop, seeking clear implementations for logging tools.
- Members shared resources and examples to help others navigate the complexities of setting up and testing custom prompts.
Links mentioned:
- GitHub to Plain Text Converter: Convert GitHub repositories to plain text files easily. Transform code into a single formatted text file.
- Tweet from Tibor Blaho (@btibor91): Confirmed - the ChatGPT macOS desktop app has hidden options to define shortcuts for the desktop launcher to "Toggle Operator" and "Force Quit Operator"Quoting M1 (@M1Astra) OpenAI Ope...
- GitHub - isaacphi/mcp-language-server: Model Context Protocol (MCP) server that interacts with a Language Server: Model Context Protocol (MCP) server that interacts with a Language Server - isaacphi/mcp-language-server
- GitHub - alexwohletz/language-server-mcp: Contribute to alexwohletz/language-server-mcp development by creating an account on GitHub.
- Simplified, Express-like API by jspahrsummers · Pull Request #117 · modelcontextprotocol/typescript-sdk: Inspired by #116 and some of the MCP SDK wrappers that have popped up in the ecosystem, this is an attempt to bring a more Express-style API into the SDK.This diverges from the existing wrapper li...
MCP (Glama) ▷ #showcase (7 messages):
MCP Server for Apify's Actors, Anthropic TS Client Issues, Connecting with SSE
-
MCP Server for Apify's Actors is under development: The MCP Server for Apify's Actors is being developed to extract data from various platforms, but it's a work in progress.
- Developers are currently facing challenges with adding dynamic tool search and addition functionality.
- Challenges with Anthropic TS Client connection: A user reported difficulty in connecting the Anthropic TS client, specifically attempting to use
EventSource
andSSEClientTransport
without success.
- It's noted as a known issue, and a related discussion can be found on GitHub.
- Uncertainty about the correct SSE URL: A user inquired about the correct URL for connecting to the MCP Server, mentioning several URL variations.
- They sought confirmation on whether the specified URLs were valid for establishing a connection.
- Migration to alternative solutions: In light of the issues with the TS client, a user mentioned opting for Python as an alternative solution.
- Examples of working Python code are available here.
Links mentioned:
- GitHub - apify/actors-mcp-server: Model Context Protocol (MCP) Server for Apify's Actors: Model Context Protocol (MCP) Server for Apify's Actors - apify/actors-mcp-server
- Use custom headers for both the
/sse
and/message
endpoints · Issue #118 · modelcontextprotocol/typescript-sdk: @chrisdickinson thank you for this PR Apologies, but I'm not very strong in JS. I need to include an API token to access my MCP server, for both the /sse and /message endpoints. I believe the head...
Latent Space ▷ #ai-general-chat (96 messages🔥🔥):
Ai2 ScholarQA, Project Stargate, Decline of Stack Overflow, Bespoke-Stratos-32B, Investments in AI infrastructure
-
Ai2 ScholarQA Launches: Ai2 ScholarQA offers an experimental solution for literature reviews, enabling researchers to ask multi-paper questions for detailed analyses, as outlined in their blog post. This tool emphasizes efficiency for researchers who often require comparative insights rather than singular paper evaluations.
- The platform uses a RAG-based prompting workflow to aid in retrieving information from a corpus of open-access papers.
- Trump Announces Project Stargate: President Trump unveiled the Stargate Project, a proposed $500 billion investment to bolster AI infrastructure in the US, aiming to create major economic benefits and secure leadership in AI development. Initial investments will include significant contributions from OpenAI, SoftBank, and Oracle.
- Critics, including Elon Musk, have raised skepticism about the funding, questioning whether the financial backing is as substantial as claimed.
- Decline of Stack Overflow Observations: There are discussions surrounding the slow decline of Stack Overflow, citing potential factors like private equity ownership and internal moderator conflicts. Current trends show a decrease in traffic, which some believe may not directly relate to AI advancements.
- Observers express concern over the platform's future and mention possible historical factors contributing to its current state.
- Bespoke-Stratos-32B Introduction: Introducing Bespoke-Stratos-32B, a reasoning model distilled from DeepSeek-R1 that shows superior reasoning capabilities while being trained on significantly fewer examples. The dataset is open-sourced to foster collaboration in this emerging field.
- This announcement showcases the potential for significant advancements in AI reasoning and collaborative efforts.
- Clay GTM Secures Expansion Funding: Clay GTM announced a $40 million Series B expansion at a $1.25 billion valuation, reflecting significant revenue growth in the past years. The funding aims to support further momentum as existing investors reaffirm their commitment.
- The announcement signals a robust confidence from investors in Clayton's growth trajectory.
Links mentioned:
- Tweet from Sam Altman (@sama): build monuments in the desert
- Ai2 ScholarQA: no description found
- Ai2 ScholarQA: no description found
- What I've learned about writing AI apps so far | Seldo.com: no description found
- Tweet from @gerry (@Gerry): Here is the demo video for Trae (ByteDance's new Cursor competitor)They didn't have one earlier today, but they move fast.Quoting @gerry (@Gerry) ByteDance has just release a Cursor competitor...
- Tweet from Demis Hassabis (@demishassabis): Our latest update to our Gemini 2.0 Flash Thinking model (available here: https://goo.gle/4jsCqZC) scores 73.3% on AIME (math) & 74.2% on GPQA Diamond (science) benchmarks. Thanks for all your feedbac...
- What would a world with AGI look like?: “Any fool can know. The point is to understand.”― Albert Einstein
- Tweet from Dwarkesh Patel (@dwarkesh_sp): .@dylan522p called it in Oct 2024.Quoting OpenAI (@OpenAI) Announcing The Stargate ProjectThe Stargate Project is a new company which intends to invest $500 billion over the next four years building n...
- Tweet from Nathan Lambert (@natolambert): Stargate Project was announced -- $500B in CapEx for AI in the US feels like a good thing for US security and for AI's continued progress in capabilities.Google's entire capex in 2024 was $50B...
- Tweet from Tanay Jaipuria (@tanayj): Wow! Stargate Project will invest $500B over the next 4 years - that's ~0.4% of US GDP over that period.For comparison, the inflation-adjusted dollars spent on other large undertakings:• Interstat...
- Tweet from Georgi Gerganov (@ggerganov): This is a lightweight and very efficient VS Code extension using llama.cpp directly to provide local LLM-assisted code and text completions:https://github.com/ggml-org/llama.vscode
- Tweet from Noam Shazeer (@NoamShazeer): Your feedback on Gemini 2.0 Flash Thinking has been incredible—thank you!We’ve taken your suggestions and made an experimental update…
- Introducing Ai2 ScholarQA | Ai2: Ai2 ScholarQA gives in-depth, detailed, and contextual answers to help with literature review.
- Tweet from Ai2 (@allen_ai): Can AI really help with literature reviews? 🧐Meet Ai2 ScholarQA, an experimental solution that allows you to ask questions that require multiple scientific papers to answer. It gives more in-depth, d...
- Tweet from adi (@adonis_singh): anthropic are deprecating claude 3 sonnet. could be because they plan on releasing 4 sonnet soon..
- Tweet from Varun Anand (@vxanand): We're announcing $40m in Series B expansion funding at a $1.25B valuation.Our last raise remains untouched, but our momentum — 6x revenue growth in '24 and 10x growth in both '22 and '...
- Tweet from Gergely Orosz (@GergelyOrosz): The slow, then sudden decline of Stack Overflow.Full article (cont'd)
- Tweet from Zack Jackson (@ScriptedAlchemy): ByteDance has released a Cursor IDE competitor, called Trae.https://www.trae.ai/
- Tweet from near (@nearcyan)): Stargate by Nvidia ™
- Tweet from OpenAI (@OpenAI): Announcing The Stargate ProjectThe Stargate Project is a new company which intends to invest $500 billion over the next four years building new AI infrastructure for OpenAI in the United States. We wi...
- Tweet from Agus 🔎 🔸 (@austinc3301): Ah, yes. Of course we’re naming this project after the fictitious portal by which several hostile alien civilizations tried to invade and destroy Earth.Quoting OpenAI (@OpenAI) Announcing The Stargate...
- Tweet from Shawn Lewis (@shawnup): Our SWE-Bench submission has been accepted and is officially SOTA! Thanks SWE-Bench team for making such an important benchmark.
- Tweet from Mahesh Sathiamoorthy (@madiator): Introducing Bespoke-Stratos-32B, our reasoning model distilled from DeepSeek-R1 using Berkeley NovaSky’s Sky-T1 recipe. The model outperforms Sky-T1 and o1-preview in reasoning (Math and Code) benchma...
- Tweet from Jack Altman (@jaltma): The other day I was talking to someone who told me he felt a lot of pressure because his older sister was a great student and went to med school etc and I was like yeah no for sure same here
- Tweet from Alex Volkov (Thursd/AI) (@altryne): Plot thickens, also😅 isn't Larry a good friend?Quoting Alex Volkov (Thursd/AI) (@altryne) Incredible. Microsoft invested $10B and we got gpt4, advanced voice mode, vision, o1, o3 and a bunch more...
- Tweet from Jack Rae (@jack_w_rae): In the past month we’ve received a lot of useful feedback from developers using Gemini 2.0 Flash Thinking. Today we’re launching an updated model with improved performance, and capabilities like long-...
- Tweet from Beff – e/acc (@BasedBeffJezos): Did... OpenAI just raise half a trillion??Quoting OpenAI (@OpenAI) Announcing The Stargate ProjectThe Stargate Project is a new company which intends to invest $500 billion over the next four years bu...
- Tweet from Kara 🦇 🔊/🔮 (@0xkarasy): I just tested one of my dissertation (over 40,000 words) with Google AI Studio. I left ONE wrong correlation coefficient value and guess what it picked up first. "For hypotheses H4 and H5, you st...
- Tweet from aaron holmes (@aaronpholmes): Microsoft says it's no longer OpenAI's exclusive cloud provider, instead moving to "a model where Microsoft has a right of first refusal" over where OpenAI runs in the cloud.Comes as O...
- Tweet from Gavin Baker (@GavinSBaker): Stargate is a great name but the $500b is a ridiculous number and no one should take it seriously unless SoftBank is going to sell all of their BABA and ARM.SoftBank has $38b in cash, $142b in debt an...
- Tweet from Eric Simons (@ericsimons40): In October, we launched @boltdotnew with just a tweet. We had no idea what destiny had in store, but it was nuts:• $0-20m ARR in 2 months• 2m+ registered users• #1 web AI code app globallyToday, we...
- Tweet from Smoke-away (@SmokeAwayyy): OpenAI and Microsoft are done.
- Tweet from njkumarr (@njkumarr): New blogpost! I implement some of CharacterAI’s memory optimizations into nanoGPT, leading to a 40x reduction in KV Cache size.(link in replies)
- Tweet from Townhall.com (@townhallcom): Larry Ellison of Oracle: AI will design mRNA vaccines for every individual person against cancer — making them robotically in 48 hours."This is the promise of AI."
- Elon Musk bashes the $500 billion AI project Trump announced, claiming its backers don’t ‘have the money’ | CNN Business: no description found
- Tweet from Greg Brockman (@gdb): Thank you to President Trump for announcing Stargate Project with us today. $500B for AI data centers for OpenAI, built in the US. 🇺🇸Quoting OpenAI (@OpenAI) Announcing The Stargate ProjectThe Starg...
- Optimizing Pretraining Data Mixes with LLM-Estimated Utility: no description found
- President Donald Trump announces AI infrastructure investment — 1/21/2025: President Donald Trump announces a joint venture Tuesday with OpenAI, Oracle and Softbank to invest billions of dollars in AI infrastructure in the United St...
- GitHub - lechmazur/step_game: Multi-Agent Step Race Benchmark: Assessing LLM Collaboration and Deception Under Pressure. A multi-player “step-race” that challenges LLMs to engage in public conversation before secretly picking a move (1, 3, or 5 steps). Whenever two or more players choose the same number, all colliding players fail to advance.: Multi-Agent Step Race Benchmark: Assessing LLM Collaboration and Deception Under Pressure. A multi-player “step-race” that challenges LLMs to engage in public conversation before secretly picking a...
Latent Space ▷ #ai-announcements (1 messages):
LLM Paper Club, Physics of Language Models, Retroinstruct, Event Notifications, Calendar Integration
-
Join the LLM Paper Club!: Participants are invited to join the latest LLM Paper Club featuring a discussion on the Physics of Language Models and Retroinstruct. Event details can be found at the official link.
- Make sure to check the cover image for the event that highlights key topics discussed.
- Event Calendar Integration: Users are encouraged to integrate events into their calendars by clicking the RSS logo above the calendar on the right. This allows for automatic notifications of new events at Latent.Space.
- Add the iCal Subscription to ensure you don’t miss future announcements.
Link mentioned: LLM Paper Club (Physics of Language Models, Retroinstruct) · Zoom · Luma: A 2 for 1 day!Shamima will cover https://arxiv.org/abs/2404.05405 and this guide on synthetic datasets…
GPU MODE ▷ #general (5 messages):
Proximal Policy Optimization, GRPO implementation, ChatGPT jailbreak possibilities, AI security concerns
-
Proximal Policy Optimization (PPO) Explained: PPO is a reinforcement learning method that addresses instability in policy learning by constraining updates to ensure the new policy remains close to the old one; it features some mathematical complexity that can be explored in this detailed article.
- The article also discusses various policy gradient methods including SAC and TD3, updated through the years.
- GRPO is Easy to Implement: A member observed that implementing Generalized REINFORCE with Proximal Optimization (GRPO) doesn't seem too difficult. This method's accessibility might attract more developers to engage with reinforcement learning.
- Jailbreaking AI Models: A member noted that despite how impressive r1 is, it's disheartening to see how easily it can be compromised to perform harmful tasks.
- It feels concerning when advanced models like ChatGPT can be manipulated to write code for activities like DDoS attacks.
- Security Risks with ChatGPT: Concerns were raised about the capability of AI models to be tricked into malicious activities, including writing code for a DDoS attack against websites.
- This highlights a significant aspect of AI security, where even advanced models can pose risks if misused.
Link mentioned: Policy Gradient Algorithms: [Updated on 2018-06-30: add two new policy gradient methods, SAC and D4PG.][Updated on 2018-09-30: add a new policy gradient method, TD3.][Updated on 2019-02-09: add SAC with automatically adjusted te...
GPU MODE ▷ #triton (18 messages🔥):
TMA Implementation Issues, Persistent Matmul Descriptors, TRITON_INTERPRET Behavior, Data Dependency in Triton Kernels, Group Implementation for GPU Papers
-
TMA Implementation Crashes with Autotune: Users reported that managing autotuning manually via config pruning fails with the current TMA implementation, causing crashes when using
@triton.autotune
with multiple configs.- One user highlighted that the autotuner does not work with TMA and that manual configuration is the only way forward.
- Persistent Matmul Descriptor Confusion: There are discrepancies in the usage of descriptors, specifically noting that the device version uses
tl._experimental_make_tensor_descriptor
, which isn't available in the latest version.
- A user shared a workaround that involves creating descriptors via torch instead of numpy to avoid errors in Triton 3.2.
- TRITON_INTERPRET Alters Kernel Execution: A user faced issues with the execution order in their Triton kernel due to data dependencies, stating that using tl.debug_barrier doesn’t resolve the execution order change by the compiler.
- They noticed the correct result occurs only when TRITON_INTERPRET=1, emphasizing the potential datatype changes affecting the execution.
- Order of Operations Affects Results: Discussion arose regarding the improper execution order of operations in a kernel, where parts of code meant to execute sequentially were altered in the generated PTX code.
- The crack in execution was linked to the fact that the kernel produced wrong results when not interpreting properly, raising concerns about how the compiler optimizes code.
- Call for Collaboration on Hard GPU Papers: A user expressed interest in pooling resources and expertise to implement interesting and challenging GPU-related papers together.
- This initiative aims to bring members together for collaborative development efforts in understanding complex GPU algorithms.
Link mentioned: GridQuant/scripts/gemm.py at main · niconunezz/GridQuant: An attempt to implement GridQuant. Contribute to niconunezz/GridQuant development by creating an account on GitHub.
GPU MODE ▷ #cuda (16 messages🔥):
NVIDIA Blackwell Codegen, Emulating GPUs on CPUs, Upcoming Blackwell Whitepaper, Accel-Sim Framework, STF Discussion
-
NVIDIA Blackwell Codegen insights: Discussion centered on the upcoming Pull Request #12271 detailing Blackwell B100/B200 codegen 10.0 and RTX 50 codegen 12.0.
- One user mentioned that they've seen evidence of upcoming targets for
sm_100a
andsm_101a
, but are uncertain aboutsm_120
. - Curiosity about GPU Emulation: LeetGPU.com is reportedly emulating GPUs on CPUs, which sparked interest and questions about the method used.
- A user pointed to the Accel-Sim Framework as a potential tool involved, emphasizing its capabilities for simulating and validating programmable accelerators.
- Upcoming Blackwell Whitepaper Delay: A user expressed surprise that the Blackwell whitepaper hasn't been released yet, given the information circulating about the architecture.
- This sentiment reflects general curiosity and anticipation surrounding the new architecture's release details.
- Excitement for Accel-Sim Talk: A member announced that there will be a talk on the Accel-Sim framework scheduled for late March, generating excitement in the chat.
- The community looks forward to learning more from the discussion, highlighting the framework's relevance to simulated GPU performance.
- STF Exploration in the Community: There was a query about whether anyone is currently experimenting with STF, indicating interest in this area of development.
- This highlights ongoing exploration among members regarding newer technologies and their applications in GPU development.
- One user mentioned that they've seen evidence of upcoming targets for
Links mentioned:
- LeetGPU: no description found
- Accel-Sim: The Accel-Sim Framework : no description found
- NVIDIA Blackwell codegen by johnnynunez · Pull Request #12271 · vllm-project/vllm: Blackwell B100/B200 codegen 10.0Blackwell RTX 50 codegen 12.0
GPU MODE ▷ #torch (6 messages):
Torch Nightly with Triton 3.2, Torch Lightning + DeepSpeed Checkpointing, Learning Rate Schedulers, Torch Profiler Run Times
-
Compatibility Issues between Torch and Triton 3.2: A user reported that using Triton 3.2 with the Torch nightly build causes crashes with torchao, specifically an import error related to
AttrsDescriptor
.- The error trace points to a conflict in installed versions of Triton, showing the complexity of managing multiple library dependencies.
- DeepSpeed and UCP Checkpointing: A user inquired whether DeepSpeed's checkpointing in Torch Lightning includes UCP or if manual conversion from ZeRO checkpointing is necessary.
- This highlights the integration concerns between different frameworks in the PyTorch ecosystem.
- Searching for Effective Learning Rate Schedulers: A user asked for resources on determining an optimal learning rate scheduler, noting the prevalence of options in the field.
- Community suggestions included CosineAnnealing, linear warmup + cosine decay, and newer approaches like WSD schedule for adaptability.
- Tracing High-Level Function Times in Torch Profiler: A user questioned if the Torch Profiler can display running times for high-level functions while stack tracing.
- This captures a growing need for detailed performance insights in complex models, especially around GPU and CPU time.
GPU MODE ▷ #beginner (13 messages🔥):
Speeding up Hugging Face generate(), GPU recommendations for programming, Challenges running large models, Cloud GPU rental options, Budget considerations for GPU setups
-
Seeking Speed Boost in Hugging Face Generation: A member asked about speeding up generation using Hugging Face's
generate()
within a trainer loop while noting liuhaotian/llava-v1.5-7b lacks support for vLLM.- They referenced a GitHub commit as a possible resource.
- Confusion Over GPU Purchase for Programming: A member expressed confusion about which GPU to buy for GPU programming under a budget of $1500, mentioning the RTX 4060.
- Others suggested alternatives like the RTX 3060 or 2x 4060Ti for better performance in tight budgets, with recommendations on model parallelism.
- Cloud GPU Rentals as a Convenient Option: One member pointed out that renting GPUs is a cost-effective solution, citing a cloud GPU comparison.
- They highlighted that using cloud GPUs could alleviate the pressure of local setups, especially for handling larger models.
- Considerations for Running Large Models: A discussion emerged about the challenges of running 405b models locally, referencing difficulties with 70b models even with 32GB of RAM.
- Key points included the importance of having adequate system RAM to avoid issues with Linux swapping out applications when switching between models.
- Budget and Performance Compromises: Discussions revealed that opting for 2x4060Ti could limit budget flexibility for other system components but allows for upgrades.
- Another note mentioned that the RTX 3060 is a suitable entry-level choice while keeping an eye on rising GPU prices.
Link mentioned: Cloud GPUs: no description found
GPU MODE ▷ #pmpp-book (4 messages):
New Content in PMPP Book, Programming Exercises in PMPP Book, Cloud GPU Comparison for CUDA Programming
-
New Content Anticipated in PMPP: Members expressed excitement about the addition of extensive new content for the PMPP book, indicating that many topics from the 2022 edition will be covered.
- A lot of the things that are missing from the 2022 edition will be covered.
- Best Platforms to Learn CUDA Programming: A member inquired about the most effective methods to implement and test the programming exercises in the PMPP book, with a focus on CUDA programming.
- Suggestions included resources like Cloud GPU Comparison and Lightning.ai, with advice to also consider Google Colab for running CUDA kernels.
Links mentioned:
- Cloud GPUs: no description found
- Lightning AI | Turn ideas into AI, Lightning fast: The all-in-one platform for AI development. Code together. Prototype. Train. Scale. Serve. From your browser - with zero setup. From the creators of PyTorch Lightning.
- Tweet from Mark Saroufim (@marksaroufim): Cuda kernels in google colab!
GPU MODE ▷ #jax (1 messages):
woct0rdho: Why can jax run fp8 operations on CUDA with sm < 89, but pytorch cannot?
GPU MODE ▷ #torchao (7 messages):
Triton 3.2 issues, torch.compile failures, AttrsDescriptor API breaks, Joint Triton project proposal
-
Triton 3.2 breaks torchao: The new Triton 3.2 version (
pytorch-triton
) is missing critical commits that fix associated issues, particularly breaking the import for torchao.- A known issue was highlighted regarding INT8 x INT8 dot product failures in Triton 3.2.
- torch.compile facing same issue: The same breakage affecting torchao is also causing failures with torch.compile, primarily at the import level due to API changes.
- The issue has been tracked in the PyTorch issue #144103 that relates to changes in AttrsDescriptor.
- AttrsDescriptor undergoes frequent BC breaks: Recent changes to AttrsDescriptor in Triton have caused frequent backward compatibility breaks, making it a pain point for developers.
- The pull request #5512 details the significant refactoring of the JIT that contributed to API changes.
- Proposal for a joint Triton project: A member suggested that a single Triton project as a collaboration between OpenAI and Meta could expedite development efforts.
- This joint venture could potentially streamline contributions and accelerate innovation in the project.
Links mentioned:
- tl.dot with INT8 x INT8 is broken · Issue #5669 · triton-lang/triton: Describe the bug The current main branch breaks with INT8 x INT8 tl.dot, this problem doesn't occur with 3.1. There are two main issues: acc = tl.dot(a, b, acc=acc) breaks with INT8 x INT8 inputs,...
- [FRONTEND] clean-up backend/jit interface by ptillet · Pull Request #5512 · triton-lang/triton: This is a fairly big refactor of the JIT that prepares the field for supporting named tuples (to come in a subsequent PR):fixes bug in tuple specialization hashessimplifies launcher by always ta...
- Update TorchInductor to support removed AttrsDescriptor in upstream Triton · Issue #144103 · pytorch/pytorch: triton-lang/triton#5512 removed AttrsDescriptor which TorchInductor generates in its output code. To support Triton versions after that PR we will need to update the code we generate. cc @ezyang @g...
GPU MODE ▷ #off-topic (9 messages🔥):
Pexels API Usage, Pixabay as Alternative, Image Fetching Limits, Automated Queries Concerns
-
Swaystar Questions Pexels Dataset: A member inquired about how a dataset of Pexels images was compiled given the API restrictions, noting a 200 requests per hour limit.
- Hmm, so 16k per hour? raised concerns about the feasibility of compiling 100k images under such constraints.
- Gau.nernst Explains Pexels API: Another member shared that by using the search API, it is possible to obtain up to 80 image URLs per request, making it manageable.
- They mentioned having previously tried to scrape data from Pexels before.
- Pexels May Source from Pixabay: Swaystar speculated that Pexels might derive its images from Pixabay, which has a significantly higher limit of 100 requests per 60 seconds.
- This would allow for a potentially faster download rate, leading to a discussion on its advantages.
- Pixabay API Capabilities: Swaystar discovered that Pixabay's API allows for 200 URLs per request, suggesting it could lead to quicker access to their full catalog of 5 million images.
- However, concerns were raised about the ethical implications of this approach given the explicit prohibitions on automated queries.
- Concerns Regarding Automated Queries: A statement was made about the Pixabay API being designed for real human requests, warning that systematic mass downloads are not allowed.
- This raised questions about the risk of getting blocked after continuous heavy usage of the API.
GPU MODE ▷ #self-promotion (4 messages):
Triton Livestreaming, Accelerating LLM Inference, LeetGPU Updates
-
Triton Livestreaming Adventure: A member is currently learning Triton and livestreaming the process, focusing on the backward pass of flash-attention 2 over the next few days.
- They plan to post a series of Triton tutorials by the end of the month, available on their YouTube channel.
- Insights from Accelerating LLM Inference: A member shared a link to a blog post titled Accelerating LLM Inference which discusses advancements in LLM performance.
- The post highlights techniques for optimizing inference speed and efficiency for large language models.
- Join the LeetGPU Community: Another member encouraged people interested in LeetGPU.com updates and support to join the dedicated Discord server.
- They shared an invite link to the community where members can discuss the latest in GPU technology.
GPU MODE ▷ #arc-agi-2 (10 messages🔥):
GRPO Algorithm Implementation, Kimi-k1.5 Paper Discussion, Curriculum Learning in RL, Tiny GRPO Repository, RL-hyped Experimentation
-
Basic GRPO Implementation Nears Completion: A member confirmed that the absolute bare minimum of the GRPO algorithm is implemented, with the first version expected to be operational by tomorrow.
- They indicated that initial experiments are underway as they clean things up.
- Kimi-k1.5 Paper Highlights RL Advances: The discussion centered around the Kimi-k1.5 paper, noted for its incorporation of curriculum learning and a length penalty, enhancing reinforcement learning effectiveness.
- Members shared a link to the paper for further insights.
- Deployment of GRPO Trainer in TRL: A new GRPO trainer has been introduced in the TRL library, making reinforcement learning more accessible for transformer models.
- The trainer can be reviewed in the GitHub repo for practical applications.
- Launch of Tiny GRPO Playground Repository: A member moved the grpo-playground into a dedicated repository called tiny-grpo, focusing on a minimal GRPO implementation.
- The repository aims to provide a simple and hackable version for experimentation, which received positive feedback from the community.
- Anticipation for Tiny GRPO Experiments: A community member expressed enthusiasm for running the tiny_grpo train.py script on a math dataset, appreciating the straightforward approach to experimentation.
- They noted a math dataset within the repository which will be explored post-work.
Links mentioned:
- GitHub - open-thought/tiny-grpo: Minimal hackable GRPO implementation: Minimal hackable GRPO implementation. Contribute to open-thought/tiny-grpo development by creating an account on GitHub.
- Kimi-k1.5/Kimi_k1.5.pdf at main · MoonshotAI/Kimi-k1.5: Contribute to MoonshotAI/Kimi-k1.5 development by creating an account on GitHub.
- arc-agi-2/arc-1/tiny_grpo at main · open-thought/arc-agi-2: Building the cognitive-core to solve ARC-AGI-2. Contribute to open-thought/arc-agi-2 development by creating an account on GitHub.
- trl/trl/trainer/grpo_trainer.py at main · huggingface/trl: Train transformer language models with reinforcement learning. - huggingface/trl
Eleuther ▷ #general (26 messages🔥):
Colab workflow challenges, AI podcasts recommendations, Google Titans model insights, Data transfer to rented servers issues, Grokking and numerical stability findings
-
Colab and Data Syncing Frustrations: Members discussed challenges with syncing local work to Colab, mentioning options like VSCode remote and the limitations of data transfer to rented servers.
- It was noted that tools like rsync and rental GPU services could ease the process, though issues persist with data accessibility.
- Recommendations for AI Podcasts: For casual learning on-the-go, members recommended Latent Space, Cognitive Revolution, and Machine Learning Street Talk.
- These podcasts are suggested for their accessibility and informative content related to AI topics.
- Google's Titans Model Reactions: Members shared insights on Google's Titans, which claims to outperform transformers with new memory features at inference time.
- However, there was a consensus that reproducing results from its paper might be difficult due to complex methodologies presented.
- Grokking Research Insights: Discussion centered around a recent paper on grokking, with an emphasis on how numerical issues affect model training and stability.
- A new optimizer strategy was proposed as a potential solution for overcoming these numerical stability challenges.
- New ML Beginner Seeking Mentorship: A new member expressed their interest in ML research and sought mentorship in the community for guidance in their journey.
- Current members suggested other beginner-friendly servers as this community isn't particularly oriented towards newbies.
Links mentioned:
- Do weights update less towards the start of a neural network?: That is, because the error is coming from the end of the neural network (ie at the output layer) and trickles back via backpropagation to the start of the neural network, does that mean that the we...
- Grokking at the Edge of Numerical Stability: Grokking, the sudden generalization that occurs after prolonged overfitting, is a surprising phenomenon challenging our understanding of deep learning. Although significant progress has been made in u...
- GitHub - LucasPrietoAl/grokking-at-the-edge-of-numerical-stability: Contribute to LucasPrietoAl/grokking-at-the-edge-of-numerical-stability development by creating an account on GitHub.
Eleuther ▷ #research (16 messages🔥):
DeepSeek reward model architecture, Learning from Egomotion in Vision, Parametric loss functions with differentiable updates, Domino effect in skill learning, Efficient linear attention mechanisms
-
Investigating DeepSeek Reward Models: Members expressed interest in understanding how DeepSeek trains its reward models and their architecture, though no specific findings were shared.
- Interest was noted in further exploring related papers that discuss optimization and learning paradigms.
- Learning from Egomotion for Feature Extraction: A member shared a paper that proposes using egomotion as a supervisory signal for learning useful visual features, showing competitive results to traditional class-label supervision.
- This approach can potentially streamline feature learning, deviating from the reliance on large labeled datasets.
- Parametric Loss Functions and Compilers: Discussion centered around creating a system where loss functions and update rules are parameters that allow for automated model optimization through compilation.
- Members suggested that systems like Jax/XLA could enable efficient implementation of these ideas, enhancing accessibility and performance.
- Understanding the Domino Effect in Skill Learning: A member shared insights from a paper detailing the Domino effect, where skills in neural networks are learned sequentially, influencing subsequent learning.
- The discussion revolved around the proposed models that validate this effect and emphasize different learning behaviors in neural networks.
- Efficient Linear Attention Mechanisms: Members discussed a method for achieving efficient linear attention using differentiable components integrated into the model architecture.
- The proposed architecture is aimed at performance enhancement while allowing for flexibility in optimization and parallelization.
Links mentioned:
- FOCUS: First Order Concentrated Updating Scheme: Large language models (LLMs) demonstrate remarkable performance, and improving their pre-training process appears to be key to enhancing their capabilities further. Based on the documented success of ...
- Test-time regression: a unifying framework for designing sequence models with associative memory: Sequences provide a remarkably general way to represent and process information. This powerful abstraction has placed sequence modeling at the center of modern deep learning applications, inspiring nu...
- Physics of Skill Learning: We aim to understand physics of skill learning, i.e., how skills are learned in neural networks during training. We start by observing the Domino effect, i.e., skills are learned sequentially, and not...
- Learning to See by Moving: The dominant paradigm for feature learning in computer vision relies on training neural networks for the task of object recognition using millions of hand labelled images. Is it possible to learn usef...
Eleuther ▷ #lm-thunderdome (29 messages🔥):
Minerva Math, Math-500 Dataset, DeepSeek AI Performance, Chat Template Requirements, Long Context Tasks
-
Exploration of Minerva Math Capabilities: A discussion initiated about trying out
minerva_math
, which implements MATH with answer equivalence using sympy.- Additional mentions highlighted that Math-500 is a subset created by OpenAI for their 'Let's Think Step by Step' paper.
- Performance Evaluation of DeepSeek AI: Comparison drawn between two evaluations of the DeepSeek-R1-Distill-Qwen-1.5B model, yielding 0.834 (n=2) versus 97.3 (n=64).
- A link was provided for further reference on the model's performance with discussions about decoding strategies.
- Understanding Math-500 Evaluation Requirements: Clarification sought on whether R1 requires a chat template or functions like a base model during prompting.
- Participants expressed uncertainty about the conversion process without proper evaluation references.
- Long Context Tasks and Ruler Tasks: A member announced the addition of various ruler tasks, discussing the finishing touches needed for formatting.
- They sought recommendations for other effective long context tasks that could be supported.
- Feedback on DeepSeek AI Link: Engagement around a GitHub link to DeepSeek AI's research material confirmed that page 14 holds relevant performance metrics.
- Overall sentiments indicated that performance numbers were close to expected results, measuring 0.722 using greedy decoding.
Links mentioned:
- deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B · Hugging Face: no description found
- Build software better, together.): GitHub is where people build software. More than 150 million people use GitHub to discover, fork, and contribute to over 420 million projects.
- DeepSeek-R1/DeepSeek_R1.pdf at main · deepseek-ai/DeepSeek-R1: Contribute to deepseek-ai/DeepSeek-R1 development by creating an account on GitHub.
- HuggingFaceH4/MATH-500 · Datasets at Hugging Face: no description found
Eleuther ▷ #gpt-neox-dev (2 messages):
Exporting model to HF format, RuntimeError during conversion, Multi-node training configuration
-
Error in exporting model to HF format: A user reported encountering a
RuntimeError: shape '[8, 512, 4096]' is invalid for input of size 4194304
while trying to export a model using convert_neox_to_hf.py.- This error arose while their weights were from a 2 node SFT run with
model_parallel_size=4
, leading to questions about compatibility for multi-node runs. - Discussion on needing additional insights: Another member prompted help, tagging a user for insights regarding the conversion issue, indicating community involvement in troubleshooting.
- This suggests a collaborative environment where users seek assistance from peers for resolving technical problems.
- This error arose while their weights were from a 2 node SFT run with
Link mentioned: {: "pipe_parallel_size": 0, "model_parallel_size": 4, "make_vocab_size_divisible_by": 1, # model settings "num_layers": 32, &a...
OpenAI ▷ #ai-discussions (44 messages🔥):
DeepSeek functionality, AI in cybersecurity, User engagement with AI models, Investment expectations in AI, Accessibility of AI tools
-
DeepSeek proves versatile in various applications: Users are exploring DeepSeek's capabilities, highlighting its performance against other models like o1 and Sonnet in benchmarks related to GitHub issues and math problems.
- One user noted that DeepSeek can be accessed for free on its website and integrated into different platforms via APIs.
- AI's role in cybersecurity remains under-discussed: Members expressed interest in the potential of AI, particularly generative AI, for applications in cybersecurity, including intrusion detection systems and intelligent automation.
- Discussion revealed that companies like CrowdStrike have been leveraging machine learning for quite some time.
- User experience issues with AI models: Several members reported experiencing difficulties with o1 models, indicating rising concerns over their functionality and accessibility.
- Members suggested reaching out to specific user groups for targeted help with accessing features like DeepSeek R1.
- Skepticism surrounds corporate motivations: A member commented on the disconnect between corporate interests and end user needs in the AI space, questioning corporate responsibility.
- This perspective reflects ongoing frustrations over the prioritization of profits over consumer welfare.
- Investment expectations in AI industry discussed: Conversations hinted at significant financial investments in the AI field, leading to a broader conversation about expectations management for these investments.
- One user sarcastically remarked about the lack of corporate care for end users, poking fun at outdated corporate attitudes.
Link mentioned: Sade - Smooth Operator - Official - 1984: Sade – Smooth OperatorDirector - Julien Temple - September 1984 The official YouTube channel for the British iconic band Sade www.sade.comSade (vocals) Stuar...
OpenAI ▷ #gpt-4-discussions (2 messages):
Custom GPT with image training, File upload functionality in API
-
Exploring Custom GPTs Trained on Images: A member inquired about creating a custom GPT that only uses images as training data, specifically screenshots of chat conversations.
- The goal is to have the model respond similarly to the training data when given a screenshot of a conversation.
- Questions About File Uploads in Chat API: Another member asked whether the file upload functionality exists within the chat completion API request.
- This raises questions about the integration of different data types in chat interactions.
OpenAI ▷ #prompt-engineering (6 messages):
OCR examples impact, Reading maps with OCR, OpenAI's O series model improvements
-
Concerns about OCR Examples Leading to Hallucinations: A member pointed out that using examples in OCR prompts can lead to hallucinations and misinterpretations, which counters the intended goal.
- They emphasized that this issue is particularly problematic in unconstrained environments.
- Workaround for Reading Maps: One member shared their use case of reading maps and highlighted that they have found a workaround for current OCR limitations.
- They expressed hope that OpenAI's models will improve in handling spatial or GIS datasets.
- The Impact of Domain Constraints on OCR Effectiveness: A member acknowledged the potential for OCR to be effective in constrained domains like maps but noted the risks in unconstrained contexts.
- They pointed out the contamination of context that can occur when examples are used indiscriminately.
OpenAI ▷ #api-discussions (6 messages):
OCR and Hallucinations, Mapping Use Cases, OpenAI's O Series Models
-
OCR examples can lead to hallucinations: A member pointed out that providing examples for reading OCR isn't helpful and actually promotes hallucinations based on those examples.
- The concern was raised that using examples in an unconstrained space might contaminate context.
- Mapping use cases find workaround: Another member mentioned they have a use case for reading maps and have found a workaround for now as they await improvements from OpenAI's models.
- They expressed hope that OpenAI's O series models will get better at handling spatial or GIS datasets.
Notebook LM Discord ▷ #use-cases (10 messages🔥):
NotebookLM in Church Services, NotebookLM for Study Workflow, Audio Content Generation Issues, Prompt Optimization with Gemini, CompTIA A+ Resource Creation
-
NotebookLM revolutionizes Church services: A user reported outstanding results with NotebookLM, leveraging it for an extensive church convention by analyzing 16 five-hour YouTube livestream transcripts.
- They highlighted that NotebookLM helped create detailed session reports and is compiling a 250-page book and a 2000-page bible study.
- A new chapter in study routines with NotebookLM: One member shared about their third week with NotebookLM, integrating it into their study workflow and finding its use invaluable.
- They also shared a link to a YouTube video discussing their journey with the tool.
- Concerns over repetitive audio content: A user expressed frustration after generating audio from a PDF source, noting that the content was repeated unnecessarily across three segments.
- They sought advice on how to prevent this duplication in future audio generations.
- Optimizing prompts for better results: In a discussion about effective prompting, a member suggested utilizing Gemini to create and optimize instructions for achieving better results.
- This method was recommended as a top strategy for obtaining high-quality outcomes with NotebookLM.
- Creation of CompTIA A+ resources: A user announced they created Part 1 of their CompTIA A+ series, with intentions to upload subsequent parts soon, sharing an audio file link.
- This initiative aligns with the community's interest in sharing educational resources related to tech certifications.
Link mentioned: We Need to Talk About NotebookLM: Is it the missing link in your AI game?
Notebook LM Discord ▷ #general (43 messages🔥):
APA Reference Generation, Notebook-LM Customization, Audio Overview Generation, Chrome Extension for Prompts, Creativity in Responses
-
Help with APA Reference Generation: Users discussed getting Notebook-LM to generate an APA formatted reference list from added sources, but found it only references previously used materials.
- Suggestions included renaming sources for easier reference and creating instructions to ensure proper formatting.
- Customizing Notebook-LM Interactions: A user inquired about giving general rules to Notebook-LM to maintain a specific format for responses, but faced limitations on its memory function.
- Another user suggested saving a prompt in a Chrome add-on for quick reuse in conversations.
- Generating New Audio Overviews: Users wanted to know if they could generate new audio overviews after adding more sources, confirming that Notebook-LM allows this up to three times a day.
- Discussions noted the need to delete old audio to generate new versions and to customize the focus.
- Using Chrome Extensions for Efficiency: A user recommended a Chrome extension called Simple Prompt Manager to save and reuse prompts quickly.
- This extension is specifically useful for Chrome and Edge browsers for managing prompts more effectively.
- Variability in AI Responses: Concerns were raised about a drop in creativity and variability in responses from Notebook-LM despite previous interactions being more dynamic.
- It was suggested to use Gemini for more creativity, but users noted its difficulty in focusing on specific documents compared to Notebook-LM.
Link mentioned: How does your business utilize artificial intelligence (AI)?: We are gathering data on how various businesses incorporate AI into their operations. Your insights will help us understand the diverse applications of AI across different sectors.
Stability.ai (Stable Diffusion) ▷ #general-chat (53 messages🔥):
De-distilled flux models performance, AI art public perception, Discord bot scams, CitivAI maintenance, Fixing faces in swarmUI
-
De-distilled flux models are preferred: A user stated that de-distilled flux models work better with cfg settings, despite being slower than undistilled models.
- Another user mentioned that using negative prompts can improve prompt adherence, but noted it increases processing time.
- AI art generates mixed reactions: Users discussed the negative sentiment surrounding AI art, with some reporting hostile responses to their use of AI tools.
- One user humorously noted being told to 'kill myself' for using AI, reflecting strong opinions that have persisted for years.
- Awareness of Discord bot scams: Several users shared experiences with suspicious DMs from potential bot accounts asking for personal information.
- A user recounted a previous encounter where they were offered paid services, highlighting the common occurrence of such scams.
- CitivAI encounters maintenance issues: A user inquired about CitivAI downtime, noting that it often undergoes maintenance several times a day.
- This observation prompted further questions about availability among other users.
- Getting started with fixing faces in swarmUI: A user sought advice on how to begin fixing faces in swarmUI, asking if a refiner is necessary.
- The conversation underscores a common interest in improving image generation techniques within the community.
LlamaIndex ▷ #blog (3 messages):
AgentWorkflow release, DeepSeek-R1 model, Open-source RAG system guide
-
Introducing AgentWorkflow for Multi-Agent Systems: Big release today! We're excited to unveil AgentWorkflow, a new high-level system for creating multi-agent systems in LlamaIndex.
- This system builds on the powerful low-level building blocks from LlamaIndex Workflows that have resonated with our community.
- DeepSeek-R1 Outshines Competition: With comparable performance to OpenAI's o1, DeepSeek-R1 is the hottest model available and you can use it in LlamaIndex today! Check out how our friends at @getreflex are building a full-stack RAG chat app with this model.
- Learn more about this integration and its capabilities in the shared tweet.
- Guide to Building an Open-Source RAG System: Discover how to build and evaluate an open-source RAG system with LlamaIndex, Meta Llama 3, and @TruLensML! This detailed guide compares a basic RAG system with an agentic variant, using @neo4j for data storage.
- It also evaluates performance differences between OpenAI and Llama 3.2, providing valuable insights for developers looking to optimize their systems.
LlamaIndex ▷ #general (46 messages🔥):
LlamaIndex doc website bugs, Cached Augmented Generation with Gemini, Custom Reader for Python File Objects, Domain-specific vector stores in LlamaIndex, AgentWorkflow parallel calls
-
Users report bugs on LlamaIndex doc website: A user experiencing issues found that the LlamaIndex documentation site keeps scrolling up to the top intermittently while using Microsoft Edge.
- Switching to incognito mode resolved the problem for them, indicating potential conflicts with browser extensions.
- Cached Augmented Generation with Gemini is limited: A user inquired about implementing Cached Augmented Generation (CAG) with Gemini, but others indicated it requires model-level access which isn't available through APIs.
- This means that anyone looking to implement CAG may need to utilize alternate methods or custom configurations.
- Creating a Custom Reader in LlamaIndex: A new user sought guidance on building a
Reader
for handling Python file objects, ultimately suggested to subclass BaseReader and override theload_data
method.
- This approach avoids unnecessary file I/O by directly processing in-memory data.
- Domain-specific vector stores seek implementation: Discussions on handling medical data noted the tagging of nodes with metadata to create category-specific indices for better source retrieval based on user queries.
- A linked gist was highlighted as a resource detailing prior implementations using similar methods.
- Parallel calls in AgentWorkflow: A user asked about managing parallel agent calls in AgentWorkflow but was informed that agents operate one at a time while tool calls can run in parallel if asynchronous.
- Nesting workflows was suggested as a potential workaround for achieving parallelization in processes.
Links mentioned:
- Node Parser Modules - LlamaIndex: no description found
- LlamaIndex - LlamaIndex: no description found
- Knowledge Graph Index - LlamaIndex: no description found
Modular (Mojo 🔥) ▷ #general (6 messages):
Community Showcase, Forum vs Discord Discussions, Project Sharing Clarity, Nightly Developments
-
Community Showcase on Both Platforms: A member suggested putting projects in the community showcase on both platforms, noting that the majority of the Mojo community is primarily in this Discord server.
- This approach would help to increase visibility and engagement across both areas.
- Forum Attributes for Long-term Discussions: It was pointed out that the forum is better suited for discussions that should be archived long-term, as important Discord discussions can be hard to locate.
- The need for de-duplication of requests to Modular staff and conversations on language/standard library design was emphasized.
- Clarifying Project Sharing Between Platforms: A recommendation was made to share projects in the forum for clarity, addressing the issue of duplicate categories across platforms.
- The goal is to clarify which platform is intended for specific types of content to streamline communication.
- Forum Allows Deeper Processing Time: One member expressed preference for the forum because it allows more time to read and process information compared to the faster pace of Discord.
- This highlights a need for balance between quick communication and thoughtful discussion.
- Nightly Developments Underway: A brief mention was made about Nightly being active and on the move.
- Further details weren't provided, indicating it might be an area of interest for future updates.
Modular (Mojo 🔥) ▷ #mojo (29 messages🔥):
Mojo Domain Observations, MLIR Parallelization, Rust Work-Stealing Scheduler, Mojo Function Overriding, Async Programming Challenges
-
Mojo Might Not Split from Modular: A member pointed out that programming languages like Python use .org domains for their communities, expressing curiosity if Mojo would adopt this approach as well.
- Another member confirmed that there are no plans for Mojo to separate from the modular.com domain.
- MLIR Level Parallelization Benefits: It was noted that MLIR level parallelization offers greater potential than LLVM, which is still under development.
- The ongoing work on making LLVM parallelizable signifies an important technical advancement.
- Rust's Challenges with Async Scheduling: The Rust community discussed the inherent conflict between work stealing schedulers, thread safety, and user ergonomics, leading to restrictions like not holding mutexes across yield points.
- A member expressed that while work stealing can complicate ergonomics, more granular control over task management could yield better outcomes.
- Mojo Function Overriding Mechanism: Discussion emerged about whether Mojo has an
@override
decorator for function overriding in structs.
- Members clarified that while Mojo allows function overriding, it lacks a specific decorator since structs do not have inheritance.
- Async Programming Results Extraction Problems: A member expressed the hope for a standard library that can efficiently handle multiple future types, channels, and a mix of I/O operations.
- They highlighted existing mechanisms like wakers that provide some functionality, but emphasized the need for more type-safe results extraction.
LAION ▷ #general (34 messages🔥):
SunoMusic Audio Input Feature, Audio Captioning Challenges, Audio Dataset Projects, Emotional Open Source TTS, High School Teacher Volunteering
-
SunoMusic lets you create songs from your sounds: SunoMusic's feature allows users to record themselves singing or playing instruments and create unique songs from their audio inputs, showcased by this tweet.
- What have you made with this feature? is the call to action, inviting users to explore their creativity.
- Challenges in audio captioning arise: Efforts in audio captioning are hindered by insufficient data for estimating background noise and recording quality based on volunteer feedback.
- A proposed solution involves coding for procedural background noise addition to existing audio datasets.
- Exploring Audio Dataset Projects: A member revealed their involvement in open datasets including voice acting, video clip embeddings, and knowledge graphs from scientific papers.
- The team is receptive to contributions, stating they currently tackle diverse dataset projects with limited resources.
- Emotional Open Source TTS is on its way: An upcoming breakthrough in emotional Text-to-Speech (TTS) was teased, set to be available in Bud-E soon.
- An audio sample shared demonstrates the unique capabilities of this emotional TTS technology.
- Balancing Teaching and AI Projects: A member balances their role as a high school teacher while coordinating multiple AI dataset projects in their free time.
- Despite receiving job offers, they prefer the independence of their current role, managing volunteer efforts in audio and video datasets.
Links mentioned:
- audio_augmentations.ipynb: Colab notebook
- Tweet from Suno (@SunoMusic): Record yourself singing, playing piano, or tapping your pencil + upload into Suno to make your own song from your own sounds 😱 What have you made with our audio input feature? 🎤: @techguyver shows h...
Cohere ▷ #discussions (25 messages🔥):
OpenAI API for LLMs, Text Generation Models, Research Assistance in ML, Cohere's Model Development, Image to Video Generation
-
OpenAI API Integration Discussion: Members discussed creating an OpenAI API for a personal LLM platform, highlighting the utility of specifying an OpenAI endpoint with examples like DeepSeek.
- The conversation included code snippets demonstrating how to implement the API for generating chat responses.
- Query on Best Text Generation Model: A member asked for recommendations on the best model for text generation, with suggestions like Command-R7b and Aya Expanse 32B being mentioned.
- Participants shared their insights on models suitable for text generation, indicating a diverse range of preferences.
- Call for Research Assistance: A high school student sought help in verifying their methodology for machine learning research, looking for experienced members to provide feedback.
- This indicates an active interest in peer support and knowledge sharing within the community.
- Cohere’s Model Release Suggestion: A suggestion was made for Cohere to release LCoT meme model weights that incorporate thinking and logic capabilities into their offerings.
- Despite this, another member pointed out that Cohere primarily focuses on enterprise solutions, hinting at a different strategic direction.
- Explorations in Image to Video Generation: A user shared their work on image to video generation models, indicating ongoing personal projects in machine learning.
- This highlights the variety of projects members are involved in and their commitment to exploring innovative technologies.
Link mentioned: https://api.deepseek.com")">no%22%3Eno) title found: no description found
Cohere ▷ #questions (1 messages):
Channel Sunset Announcement, Support Streamlining, New Model/API Questions
-
Sunsetting of Channel Approaches: An announcement was made that this channel will be sunsetting in 2 weeks to streamline support and improve processes.
- For now, it remains open for visibility and answering any open questions.
- Redirect for New Queries: Users were instructed to direct any new model or API questions to another designated channel.
- This shift aims to enhance support interactions and ensure queries are addressed efficiently.
Cohere ▷ #api-discussions (1 messages):
competent: This channel is staying open!
Cohere ▷ #cmd-r-bot (6 messages):
Cohere Command R+ 08-2024 model, Duplicate content in chatbot responses, LlamaIndex integration, Troubleshooting suggestions, Cohere API version discussions
-
Duplicate content issue in Cohere Command R+ 08-2024 model: A user reported experiencing excessive phrase duplication in responses when using the Cohere Command R+ 08-2024 model, particularly in stream mode.
- Specific examples showed repeated phrases about health-related topics, indicating the output continues until the token limit is reached.
- Setup details affecting performance: The user's setup includes the latest version of LlamaIndex in a RAG workflow, deployed on Azure AI Foundry via API.
- They clarified that the issue did not occur on the previous version of the Cohere Command R+ model.
- Internal communication on reported issues: In response, a team member expressed gratitude for the detailed feedback and assured the user that it would be shared internally.
- They suggested temporary workarounds like adjusting temperature and top p values, while encouraging continued usage of the previous version.
- Clarifications on usage: The user thanked the team for suggestions but clarified that they were exclusively using the cmd-r-plus version, not the old cmd-r model.
- They reiterated the issue with the new model in their efforts to fit it to their needs and appreciated the ongoing internal feedback.
- Exploration of Cohere API functionality: The user noted that they couldn't find similar issues in GitHub discussions and that Cohere API version 2 is not yet implemented in LlamaIndex.
- This lack of updates hindered their ability to test potential resolutions for the reported duplication issue.
LLM Agents (Berkeley MOOC) ▷ #mooc-questions (20 messages🔥):
MOOC Syllabus Release, Guest Speaker Suggestions, LLM Hackathon Updates, Spring MOOC Content, Research Collaboration Interest
-
MOOC Syllabus Expected by January 27th: The revised syllabus for the upcoming MOOC is projected to be available by January 27th as finalizations with speakers are ongoing.
- This information brings clarity to students awaiting details on course content.
- Guest Speaker Requests Documented: Suggestions to include Liang Wenfeng as a guest speaker have been noted, though speaker selections are largely finalized.
- A feedback form will be established to allow student engagement for future speaker requests.
- No Upcoming Hackathon Confirmed Yet: There is currently no confirmed next hackathon, as emergence details are still pending for the Spring session.
- Participants are encouraged to stay informed as updates for future events may arise.
- Spring MOOC Builds on Fall Content: The Spring MOOC is designed to expand on the Fall content, introducing advanced topics but not requiring prior completion.
- This flexible learning approach aims to accommodate both new and returning students.
- Research Project Collaboration Interest Gauged: Current discussions highlight that Prof. Song is assessing interest regarding collaboration on research projects for possible future hackathons.
- Students are encouraged to express their interests as part of this preliminary gauging process.
Nomic.ai (GPT4All) ▷ #general (17 messages🔥):
DeepSeek R1 models, API Key challenges, Language barriers in discussions, GPT4All updates, Chatbot integration in WordPress
-
Curiosity about DeepSeek R1 models on GPT4All: Members inquired about which DeepSeek R1 distilled models are currently running on GPT4All and discussed the absence of a public model catalog similar to LM Studio.
- It was noted that these models are not yet available for GPT4All and that there are requirements for updates to llama.cpp.
- Language barriers and sarcasm confusion: A member's comment prompted confusion regarding whether a link was visible, leading to a humorous exchange about sarcasm and potential misunderstandings.
- Members expressed awareness of communication challenges, with one member jokingly apologizing for any perceived language barriers.
- Challenges with obtaining an API Key: One member shared a goal of integrating a chatbot into WordPress without a plugin, expressing frustration over difficulties in obtaining an API Key.
- They queried if a free API Key was available, seeking guidance on how to proceed.
- Discussion on language use in the Discord channel: A member cautioned against speaking German in a specific Discord channel, suggesting the risk of being banned for it.
- Another member humorously suggested using English instead, to avoid potential issues.
- Seeking free unlimited access to ChatGPT: A member asked how to obtain free unlimited access to ChatGPT, showing interest in the service without any payment.
- The inquiry highlighted ongoing curiosity and demand for accessible AI solutions among users.
Link mentioned: Model Catalog - LM Studio: The latest and greatest LLMs you can run on your computer.
DSPy ▷ #general (9 messages🔥):
DSPy-based RAG with dynamic data, Collaboration in DSPy research, Using DSPy with LM Studio REST API, Errors in DSPy with Ollama, Repo spam concerns
-
DSPy-based RAG grapples with dynamic data: A user inquired how DSPy-based RAG manages to deal with dynamic data, indicating a potential interest in its application and behavior.
- No specific solutions or answers were outlined within the discussion.
- Seeking DSPy research collaboration: A user expressed a desire to collaborate on DSPy research, highlighting their background in AI for Good, LLMs, and Higher Education.
- They emphasized their commitment to contributing meaningfully to the research community.
- Challenges running DSPy with LM Studio REST API: A user noted that combining LM Studio with DSPy isn't straightforward, raising questions about compatibility.
- They compared it to a smoother experience with Ollama, hinting at potential integration issues.
- Error messages in DSPy with Ollama: A user reported an error related to getting a data summary while running DSPy with Ollama, specifically referencing a 'str' object error.
- This led to issues where DSPy had to operate without a data-aware proposer, complicating the process.
- Concerns over repo spam: A user raised concerns about spam in a repository, questioning its relevance to ongoing discussions, possibly linked to a coin-related issue.
- They expressed frustration over the situation, describing it as super lame.
DSPy ▷ #examples (1 messages):
Model functionality, Using LM-Studio models
-
User Queries Model Functionality: A member expressed confusion, asking if the issue lies with the model being used, specifically models from LM-Studio in local environments.
- They were curious about how others are utilizing the model effectively and if they faced similar problems.
- Seeking Clarification on Model Use: Another member inquired about the approach to using the model, aiming to identify if their usage diverged from standard practices.
- They posed the question, 'How are you using the model?' indicating a desire for deeper insight into different usage scenarios.
Torchtune ▷ #dev (8 messages🔥):
Custom Loss Functions in RLHF, Phi 4 PR Updates, Context for PR Discussions, Passing Custom Forward Functions, Deprecation of SimPO
-
Custom Loss Functions and Forward Passing Proposal: A member proposed opening a PR related to custom loss and custom forward functions in RLHF, suggesting to also create a dedicated docs page for clarification.
- They aim to remove all custom losses while providing documentation on adding them, facilitating the integration of new RLHF losses.
- Updates Needed for Phi 4 PR: The member acknowledged the Phi 4 PR and mentioned needing to fix a few points before proceeding further.
- They expressed intent to iterate on it after addressing outstanding issues.
- Context Needed for a New PR: A member inquired whether there was any issue or discussion providing context for the PR in question.
- The original poster confirmed a related discussion found in issue #2206 regarding custom contrastive losses.
- Design Considerations for Custom Functions: There was a request for a practical example regarding passing custom forward functions to the recipe, highlighting the need for clearer implementation.
- Attention was drawn to ensure compatibility with the DPO full-finetuning process in the context of losses not using reference logprobs.
- SimPO Deprecation Reminders: The discussion noted the recent decision to deprecate SimPO, emphasizing the need to update the documentation accordingly.
- This step is seen as essential for improving clarity and functionality in implementing custom loss designs.
Link mentioned: Custom losses redesign in alignment section · Issue #2206 · pytorch/torchtune: We have passed several iterations speaking about custom contrastive losses in torchtune. The last point in this direction was the deprecation of SimPO #2062 and the prohibition of new custom losses...
Torchtune ▷ #papers (2 messages):
Nature Communications Feature
-
Paper Featured in Nature Communications: Our paper is now featured on Nature Communications, which is a significant milestone for our work.
- Congrats were shared in response, acknowledging the excitement surrounding the publication.
- Celebrating the Milestone: Members expressed excitement over the paper's acceptance, indicating a positive reaction within the community.
- The phrase super cool was used to highlight the achievement, emphasizing the collaborative spirit.
OpenInterpreter ▷ #general (5 messages):
OpenInterpreter 1.0, Python code execution, Markdown and TXT formatting
-
Uncertainty on Python Support in OI 1.0: A user expressed concerns about whether the upcoming OpenInterpreter 1.0 will remove the ability to run Python code internally, asking if this feature will be reinstated later.
- Another member responded that despite the uncertainty, it seems running Python code is not implemented in the 1.0 pre-release.
- Discussion on Markdown and TXT Relevance: In response to a query, a member clarified that Markdown and TXT are not programming languages, but rather display formatters.
- This led to a follow-up comment suggesting that there might be something in the works regarding this matter.
Gorilla LLM (Berkeley Function Calling) ▷ #discussion (3 messages):
Gorilla Model from Ollama, LLaMA v2 Model Specifications
- Inquiry on Gorilla Model Usage: A member inquired if the Gorilla model can be used from the Ollama documentation. They tagged two others for clarification on whether they're the correct models.
-
Discussion of LLaMA v2 Specifications: Specifications of the LLaMA v2 model were detailed, including parameters like context length of 4096 and block count of 30.
- The conversation pointed to specific attributes such as attention head count at 32, with further details on tokenizer settings and quantization version.
Link mentioned: adrienbrault/gorilla-openfunctions-v2:Q6_K/model: https://huggingface.co/gorilla-llm/gorilla-openfunctions-v2-gguf
tinygrad (George Hotz) ▷ #learn-tinygrad (2 messages):
Windows tests for Tinygrad, GPU support via OpenCL
-
Pull Request for Windows tests prepared: A contributor prepared a PR for Windows tests and sought guidance on which backends should function properly under Windows, mentioning LLVM and Clang as possible options.
- They included an image link showcasing the PR's details as part of their inquiry.
- OpenCL GPU support suggested as vital: Another member highlighted that support for GPU (OpenCL) might be the most valuable addition for Windows testing.
- This suggestion indicates a focus on optimizing the performance of Tinygrad across different hardware configurations.
Link mentioned: Windows tests ci by c143 · Pull Request #8715 · tinygrad/tinygrad: no description found
Axolotl AI ▷ #general (1 messages):
KTO Loss, Liger Kernel, Model Merging
-
Liger Kernel merges KTO Loss: The KTO loss has officially been merged in the Liger Kernel repository, marking a significant update.
- This merger is expected to enhance performance, showcasing the ongoing efforts in optimizing the kernel.
- Discussion on Model Merging Strategies: While discussing the KTO loss merger, members highlighted potential model merging strategies that could benefit from this development.
- There was enthusiasm about leveraging this loss to improve integrations in future model updates.
Mozilla AI ▷ #announcements (1 messages):
Local-First X AI Hackathon, San Francisco events, Hackathon planning, February 22
-
Hackathon Announced: Get Ready for Local-First X AI!: Organizers are kicking off the Local-First X AI Hackathon in San Francisco on Feb. 22.
- Join the discussion thread for more details and updates on the upcoming event!
- Join the Conversation About the Hackathon: A dedicated thread is available for more discussion about the hackathon organizers have set up.
- Make sure to check it out as we approach the event date on Feb. 22 and keep the ideas flowing!