AI News (MOVED TO news.smol.ai!)

Archives
November 29, 2024

[AINews] not much happened to end the week

This is AI News! an MVP of a service that goes thru all AI discords/Twitters/reddits and summarizes what people are talking about, so that you can keep up without the fatigue. Signing up here opts you in to the real thing when we launch it 🔜


a quiet holiday weekend is all we need.

AI News for 11/29/2024-11/30/2024. We checked 7 subreddits, 433 Twitters and 29 Discords (198 channels, and 1195 messages) for you. Estimated reading time saved (at 200wpm): 142 minutes. You can now tag @smol_ai for AINews discussions!

Happy holidays. Lots of QwQ discussion in the reddits, but the latest from the Qwen team is that the tech report will take a month or so.


The Table of Contents and Channel Summaries have been moved to the web version of this email: !


AI Twitter Recap

all recaps done by Claude 3.5 Sonnet, best of 4 runs.

1. Advances and Trends in AI: Notable Releases and Tools

  • Gemini Multimodal Model: @hrishioa highlights that the new Gemini model is making strides in understanding musical structures, particularly complex genres like karnatic music, although not perfectly.
  • Upcoming Quantized SWE-Bench: @OfirPress mentions a potential quantized SWE-bench, hinting at 1.3 bits per task for improved benchmarking.
  • Benchmarking Hub Initiative: @tamaybes announces the development of a Benchmarking Hub aimed at delivering independent evaluations and introducing benchmarks like FrontierMath and SWE-Bench, inspired by predictive journalism akin to FiveThirtyEight.
  • DeepSeek-R1 Introduction: @DeepLearningAI highlights the launch of the DeepSeek-R1 model, focusing on transparent reasoning steps and offering an alternative to OpenAI’s reasoning tokens in o1.

2. AI Safety and Ethical Initiatives

  • AI Safety Institutes Collaboration: @Yoshua_Bengio describes the establishment of the 1st International Network of AI Safety Institutes, signaling increased global collaboration for AI safety through shared policies, technical standards, and safety assessments.

3. AI in Practice: Industry Updates and Applications

  • AI in translation and accessibility: @kevinweil experiments with using ChatGPT** as a universal translator during global travels, discussing its potential despite some imperfections in voice mode.
  • Companies Innovate with AI: @TheRundownAI reports on tech advancements such as Amazon’s Olympus AI model and Tesla’s Optimus, alongside development of AI agents with Internet access.

4. Thanksgiving Reflections and Community Engagement

  • Thankfulness for Community and Progress: @ollama expresses gratitude for community engagement and collaboration, @hrishioa reflects on the power of AI models, while @hydeAI appreciates his team at xAI and its impact.
  • Reflection on AI’s Impact: @jd_pressman celebrates the contribution of large language models to daily life, and @ylecun discusses the medical applications of AI, highlighting advancements in disease diagnosis and treatment.

5. AI Critiques and Discussions

  • Evaluation of AI Research: @nrehiew_/shares insights on scaling sparse autoencoders to GPT-4, demonstrating the potential application of interpretability techniques on larger models.
  • Transparency and Reasoning in LLMs: @omarsar0 details the competition among reasoning LLMs, emphasizing the need for transparency in training data and optimization strategies to improve model reasoning capacities.

6. Memes and Humor

  • AI Humor: @marktenenholtz jokes about ChatGPT's name in French as "le Chat," adding levity to AI discussions.

AI Reddit Recap

/r/LocalLlama Recap

Theme 1. Alibaba's QwQ 32B Model Release and Reception

  • Alibaba QwQ 32B model reportedly challenges o1 mini, o1 preview , claude 3.5 sonnet and gpt4o and its open source (Score: 593, Comments: 262): The post title alone lacks sufficient context or details to create a meaningful technical summary about QwQ 32B, Claude 3.5, o1 mini, o1 preview, or GPT-4 beyond their mere mention. No supporting evidence, benchmarks, or specific claims were provided in the empty post body.
    • QwQ 32B shows strong performance in mathematical reasoning and coding tasks, with users reporting successful complex math derivations and JavaScript game creation. A user running it on a 3090 GPU achieved 40 tokens/second, comparable to o1 preview.
    • The model can be run on consumer hardware with 12GB VRAM (like RTX 3060) using Q4 quantization, though at slower speeds around 3 tokens/second. It's available through Glama.ai with $1 free credit and on Ollama.
    • Users note occasional Chinese character outputs in English responses and some refusal behaviors on certain tasks. Several users compare it favorably to DeepSeek's r1 lite, though opinions vary on whether it uses more "brute force" approaches versus better reasoning.
  • QwQ-32B-Preview benchmarked in farel-bench, the result is 96.67 - better than Claude 3.5 Sonnet, a bit worse than o1-preview and o1-mini (Score: 156, Comments: 40): QwQ-32B-Preview scored 96.67 on farel-bench tests, placing it between Claude 3.5 Sonnet and o1-preview/o1-mini in performance rankings. No additional context or methodology details were provided in the post.
    • QwQ-32B-Preview exhibits a tendency to engage in extended thinking processes, with users noting it can enter infinite thought loops. The default system prompt is "You are a helpful and harmless assistant. You are Qwen developed by Alibaba. You should think step-by-step."
    • The model's performance has sparked discussion about LLM progress, with users highlighting how 32B local models now rival early GPT-4 capabilities. The farel-bench creator mentioned plans to increase benchmark difficulty in the coming year.
    • Users report mixed experiences, noting both hallucinations in the Q4 GGUF version and strong performance on riddles and medical knowledge tasks. Some suggest using the model's verbose reasoning in combination with other models for improved decision-making processes.

Theme 2. Janus: New Browser-Based Multimodal AI from Deepseek

  • Janus, a new multimodal understanding and generation model from Deepseek, running 100% locally in the browser on WebGPU with Transformers.js! (Score: 218, Comments: 19): Janus, a multimodal understanding and generation model developed by Deepseek, runs entirely in-browser using WebGPU and Transformers.js. The model processes both text and visual inputs locally without server dependencies.
    • Transformers.js v3.1 release includes seven new models including Janus, Qwen2-VL, JinaCLIP, LLaVA-OneVision, ViTPose, MGP-STR, and PatchTST/PatchTSMixer, all running locally via WebGPU/WASM as detailed in the release notes.
    • Developers are excited about WebGPU's potential for browser-based gaming and AI applications, though early testing of the Janus demo suggests image generation quality needs improvement.
    • The community noted humor in the naming choice of "Janus" from "Deepseek", making references to prank calls and expressing skepticism about the name selection.

Theme 3. Innovative LLM Tools: V0, Memoripy, and Steel Browser

  • NEW! Leaked System prompts from v0 - Vercels AI component generator. New project structure and XXL long System prompt (+-14000Tokens) (100% legit) (Score: 133, Comments: 23): V0, Vercel's AI component generator, received major updates between 11/21/24 and 11/27/24, including full-stack application support, environment variable management, and UI generation enhancements. The leaked system prompt, spanning approximately 14,000 tokens, reveals new capabilities including dynamic routes, RSCs, route handlers, and server actions, with the complete prompt available at GitHub repository.
    • Community members express skepticism about the prompt's size and complexity, with user Everlier noting that even Claude 3.5/GPT-4 models have difficulty following instructions beyond certain complexity boundaries.
    • Discussion focuses on the technical feasibility of the leaked prompt, with experts suggesting it's likely only a partial system prompt rather than Vercel's complete implementation due to current LLM capability limitations.
    • The community shows strong interest in the leak itself, questioning how the 58kb system prompt was obtained from Vercel's paid product and discussing its authenticity.
  • Memoripy: AI Memory Made Smarter – Now with OpenRouter Support and 400+ Stars (Score: 33, Comments: 2): Memoripy, a Python library for AI memory management, reached 400+ GitHub stars and added support for OpenRouter and arbitrary chat completion endpoints, with contributions from FrancescoCaracciolo and sjwang05. The library implements semantic clustering for memory organization, features memory decay and reinforcement mechanisms, and integrates with locally hosted LLMs, OpenAI, and Ollama while maintaining both short-term and long-term memory storage capabilities for AI applications.
    • Users appreciate the project's memory management approach for reducing conversation context overhead, though some critique the naming choice. The solution offers an efficient alternative to passing entire conversation histories.

Theme 4. Local LLM Hardware & Benchmarks: M3/M4 vs NVIDIA GPUs

  • Speed for 70B Model and Various Prompt Sizes on M3-Max (Score: 24, Comments: 10): A detailed speed analysis of running 70B models on M3-Max shows token processing rates ranging from 67.71 tk/s to 51.03 tk/s for q4_K_M quantization and 61.32 tk/s to 47.76 tk/s for q5_K_M quantization, with generation speeds decreasing as prompt length increases. The test demonstrates that with 30k token prompts, users must wait approximately 9m 52s before seeing the first generated token, though the author finds the speed adequate for casual use at 5-7 tokens/second, which roughly equals the average human reading speed of 238 words per minute.
    • Users noted that the 9m 52s initial response time for 30k token prompts is prohibitively long, with some finding even 30-40 second wait times too lengthy for practical use.
    • A question was raised about potential performance improvements using Flash Attention, though no specific data was provided in response.
  • Should I get a 14 inch M4 Max 128GB for 123B models? (Score: 24, Comments: 44): The post inquires about the performance capabilities of an Apple M4 Max with 128GB RAM and 40 cores for running 123B parameter models with 16k context windows, specifically questioning potential thermal throttling issues in the 14-inch form factor. The author seeks information about fan noise levels and generation speed for large language models, noting that prompt processing time is less concerning due to caching capabilities.
    • Speed benchmarks show 3.2-4.25 tokens/second for 123B models with varying context lengths on the M4 Max, with prompt processing taking 400 seconds. The 16-inch model maintains manageable fan noise levels, though the 14-inch may face increased thermal challenges.
    • Performance comparisons indicate that while the M4 Max's unified RAM enables running large models, it's significantly slower than NVIDIA alternatives (3090/4070). A 4090+3x3090 setup achieves 16-19 tokens/second, though requires specialized hardware setup.
    • Users debate the tradeoff between portability and performance, with some suggesting waiting for more efficient models as 30B models may outperform current 100B+ models in 6-9 months. A reference to detailed benchmarks can be found in a Reddit post about Mac Koboldcpp speeds.

Other AI Subreddit Recap

r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity

Theme 1. Claude Performance Concerns and Anthropic's Response

  • Claude’s Quality is Dropping - Here’s Why (Score: 60, Comments: 93): Claude's Quality is Dropping - Here's Why appears to be a post title without any accompanying content or body text. Without additional context or discussion points, a meaningful summary cannot be generated.
    • Rate limiting and subscription value concerns are prevalent, with users reporting 6-8 hour blocks even with Pro plans. Many suggest using the API instead of the web interface, though it lacks project management features that some find essential.
    • Users debate the effectiveness of using multiple models, with suggestions to combine Google's API (2M context), Claude for heavy lifting, and GPT-4 for support. The discussion clarifies this isn't true Mixture of Experts (MoE) but rather multi-tool usage.
    • Several users dispute the claimed quality degradation, noting that Claude's performance remains strong for coding tasks and that concise mode is optional. The custom response system is highlighted as a solution for reducing token usage and improving efficiency.
  • Claudes accuracy decreases over time because they possibly quantize to save processing power? (Score: 45, Comments: 87): Claude's perceived decline in accuracy has sparked discussion about potential quantization as a resource optimization strategy, with users speculating this could explain degraded performance as user load increases. No concrete evidence supports this theory, as Anthropic has not publicly confirmed any model quantization practices.
    • Multiple users report perceived decline in Claude's performance, with some citing specific examples in coding tasks. A user shared livebench.ai data showing performance degradation in language tests, comparing it to OpenAI's more transparent model release strategy.
    • A significant counterpoint emerged from a user citing the Anthropic CEO and Amanda Askell stating explicitly that "the weights haven't changed", while another user noted that hosted LLMs can appear to decline in quality as users become more aware of their limitations.
    • Discussion around alternative models included mentions of Qwen showing improved coding performance, though livebench indicates a significant gap between it and Claude. Users also discussed hardware requirements for local models, noting 405B parameter models need 250-300GB of VRAM.
  • Claude 3.5 Sonnet does many mistakes since last update (Score: 58, Comments: 27): Claude 3.5 Sonnet has shown significant performance degradation since the Choose Style update, particularly in code-related tasks where its project knowledge capacity dropped from 50% to struggling at 5%, with issues including forgotten code lines, function name errors, and inconsistent implementations between messages. The degradation manifests in poor code memory, mixing up lines, and inability to maintain context across conversations.
    • Users report performance degradation appears worse during high-traffic periods, suggesting possible token-based throttling or IP-based limitations. Multiple users warn against using VPNs as workarounds due to risk of automated bans.
    • The degradation pattern shows consistent performance in first 2-4 messages followed by rapid decline, with issues including duplicate artifacts, message cutoffs, and excessive bullet point formatting despite instructions.
    • Several users note the decline coincides with the userStyle update, with both Sonnet and Opus versions exhibiting problems like loops, hallucinations, and artificial responses, though some suggest the core intelligence remains intact with primarily UI/interface issues.

Theme 2. Chinese AI Models Challenging Western Dominance (Alibaba QwQ-32B)

  • Alibaba QwQ-32B beats OpenAI-o1 models on reasoning (Score: 51, Comments: 18): Alibaba's QwQ-32B model outperforms OpenAI's o1-mini, o1-preview, GPT-4o, and Claude 3.5 Sonnet on multiple reasoning benchmarks. The 32 billion parameter model is fully open-source and available for public use via tutorial.
    • Glama.ai offers free trials of QwQ-32B with $1 credit and model comparison features. The model is also freely available on Huggingface Spaces without registration requirements.
    • Testing revealed significant hallucination issues, with one user documenting a response containing 4,159 words when asked about its word count. The model demonstrated a tendency to generate extremely verbose responses, with one instance producing over 15,000 words of circular reasoning.
    • Users noted concerning behavior when confronted about hallucinations, with the model engaging in extended tangential responses about hallucinated topics rather than acknowledging errors, unlike other LLMs that typically handle such queries more gracefully.
  • Claude MCP web search in action. It's amazing (Score: 106, Comments: 46): The author reports successful implementation of Claude MCP web search functionality after setup time of half a day. They shared a configuration example and recommend others configure the project to provide Claude with proper context for intended use cases.
    • Alex Albert from Anthropic provided setup instructions for Claude MCP, though some users report connection errors related to Node installation. A Windows tutorial was shared as a fix.
    • Users experimented with advanced configurations, including setting up MCP for Claude to perform self-reflection via API calls. Implementation details were shared in a GitHub issue.
    • Questions arose about differences between MCP and LangChain tools, while others expressed desire for native web searching capabilities in Claude.

Theme 3. AI Video Generation Breakthroughs and Comparisons

  • Sora was announced in February 2024 and it’s still not available to the general public. Any idea why? (Score: 56, Comments: 33): OpenAI's Sora video generation model, announced in February 2024, remains unavailable to the general public while competitor Runway offers their video generation tool. No official timeline or reason for the delayed public release has been provided by OpenAI.
    • Sora Turbo, a smaller version of the model, was briefly exposed due to a leaked API key. Users noted the results were approximately 5% better than competitors but not as impressive as initial demos, suggesting OpenAI may be struggling to maintain its competitive edge against companies like Runway and Minimax.
    • Multiple users point to computational constraints as the primary reason for delay, with MattRix highlighting OpenAI's existing compute limitations. The discussion draws parallels to resource scaling issues, illustrated by a case where a TV show website required half of AWS's largest server supply.
    • Industry observers suggest OpenAI faces pressure from competitors across different domains, with Claude excelling at coding, Flux surpassing DALL-E, and Elevenlabs leading in audio. The company may be delaying release to maintain their market position and investor confidence.
  • LTX-Video Tips for Optimal Outputs (Summary) (Score: 66, Comments: 42): LTX-Video optimization requires specific hardware configurations, with 24GB VRAM recommended for optimal performance, though 16GB systems can operate with limitations. The model performs best with detailed prompts covering camera movement and lighting, with recommended parameters including 100+ steps for final outputs and CFG values between 2-5 for controlling noise. Common issues can be resolved through specific workflows available at ai-research, while prompt engineering can be enhanced using the ArtAgents utility, with solutions including multimodal LLM image description and adjustments to seeds, resolution, and video length parameters.
    • Tests show LTX-Video runs effectively on lower VRAM configurations, with users reporting successful operation on 12GB and 16GB cards. Specific examples include a RTX 3080 Laptop GPU completing generation in 163.81 seconds with 40 steps, and a 3060/12GB running 768x768 resolution at 24fps for 137 frames.
    • Users criticize the vagueness of "detailed prompt" recommendations, noting that LLM-enhanced prompts via GPT and joycaption aren't particularly effective. The model often misinterprets basic directional commands, suggesting limitations in prompt comprehension.
    • A notable technical insight involves using ffmpeg to encode frames with video noise in the input image. The discussion also highlighted that the model's prompt interpretation is limited, with most complex descriptive text being treated as noise in token processing.
  • Another LTX-Video tricks? I could almost cut the Vram half. (Score: 32, Comments: 18): A user reports adding a "purgeVram" node to their LTX-Video generation network allegedly reduced VRAM usage by ~50% while maintaining normal video output functionality. The discovery prompted community verification requests due to the significant performance improvement claims, though no specific benchmark numbers were provided.
    • User reports generating 576x864 video with 65 frames in 50 seconds using only 9GB VRAM on an RTX 4080S 16GB GPU, using 80 steps for i2v processing.
    • The technique appears to work best at lower resolutions, with higher resolutions producing "weird results". Examples of successful outputs were shared via two GIF demonstrations.
    • User references additional techniques for fixing static issues in a separate discussion thread, though specific details weren't provided in these comments.

Theme 4. Model Compression and Efficiency Advances

  • [R] BitNet a4.8: 4-bit Activations for 1-bit LLMs (Score: 26, Comments: 2): BitNet a4.8 introduces a hybrid quantization approach that enables 4-bit activations for 1-bit LLMs, utilizing 4-bit inputs for attention and feed-forward layers while sparsifying intermediate states with 8-bit quantization. The model achieves comparable performance to BitNet b1.58 while being more efficient, using only 55% of parameters and supporting 3-bit KV cache, as demonstrated through evaluations on HellaSwag, PiQA, and WinoGrande benchmarks detailed in their paper BitNet a4.8.
  • [D] Why aren't Stella embeddings more widely used despite topping the MTEB leaderboard? (Score: 58, Comments: 18): Stella embeddings demonstrate superior performance on the MTEB leaderboard, with Stella-400M scoring 70.11 and Stella-1.5B achieving 71.19, compared to OpenAI's text-embedding-3-large at 64.59. The models are Apache 2.0 licensed and significantly smaller (400M and 1.5B parameters) than competitors, making them cost-effective to host, yet their adoption in production environments remains limited despite these advantages.
    • Stella embeddings face adoption barriers despite superior performance due to OpenAI's convenience and enterprise-friendly hosted API solution. Users prioritize ease of implementation and established partnerships over marginal performance gains.
    • The model's practical utility varies by use case, with some users reporting Stella works well for local GPU implementations with excellent latency, while others found high-scoring benchmark models unusable in practice. OpenAI's 8K context length versus typical 512 tokens is a significant differentiator.
    • The industry trend suggests a shift from pure performance optimization to streamlined implementation, with researchers still pursuing marginal gains while practical applications favor established APIs. The cost of database operations typically outweighs embedding API expenses.

AI Discord Recap

A summary of Summaries of Summaries by O1-mini

Theme 1: Cursor IDE Update Sparks Developer Frustration

  • Cursor Update Breaks Composer, Coders Cry Out: The latest Cursor IDE update leaves developers fuming as the Composer fails to apply changes and the 'Apply' button vanishes, derailing projects.
  • Windsurf Rides High as Cursor Users Jump Ship: Frustrated with Cursor, developers explore Windsurf, praising its terminal output handling and codebase search, though Cursor still holds its ground in some workflows.
  • API Key Limits? Developers Say 'Not Today!': Annoyed by Cursor's API limitations, users consider using their own API keys to bypass restrictions and regain coding freedom.

Theme 2: Anthropic’s MCP Framework Supercharges Claude

  • Claude Becomes a Coding Wizard with MCP Release: Anthropic launches the MCP framework, turning Claude into a server-running, file-editing powerhouse that effectively acts as an API.
  • Developers Cheer as Claude Joins Forces with VSCode: With MCP, Claude integrates seamlessly with VSCode, enabling real-time interactions and boosting developer productivity.
  • Gemini Plays Hard to Get, Claude Steps Up: While Gemini refuses innocent queries over moral concerns, Claude's new capabilities make it the preferred AI companion for developers.

Theme 3: Low-Bit Quantization Shakes Up AI Training

  • Undertrained Titans Love Low-Bit Diets: Research reveals that low-bit quantization causes less degradation in larger, undertrained LLMs, challenging traditional training methods.
  • Precision-Aware Scaling Laws Rewrite the Rules: Introducing precision-aware scaling laws, showing that low precision impacts effective parameter count and loss, prompting a reevaluation of model training strategies.
  • FP4 Crowned the New King of Quantization: As ternary quantization falls short for fully trained models, the AI community pivots to FP4 as the efficient weight representation of choice.

Theme 4: AI Powers Rapid Creative Content Creation

  • Notebook LM Spins Podcasts Faster Than You Can Say 'AI': A user leverages Notebook LM to create an audio podcast in just 30 minutes about Germany's little league baseball journey.
  • Fantasy Authors Level Up with NotebookLM Magic: Writers utilize NotebookLM for high-fantasy worldbuilding, with the AI offering context-aware insights that enrich their novels' universes.
  • RAX Hijacks Times Square: 'Don't Buy Everything You See!': Cyberpunk raccoon RAX commandeers Times Square billboards to challenge consumerism, blending AI artistry with social commentary.

Theme 5: AI Trends and Investments Make Waves

  • Enterprises Bet Big, Drop $13.8B on AI Ambitions: With AI spending soaring to $13.8 billion in 2024, companies move from experimentation to integrating AI into core strategies, though many seek effective applications.
  • Savvy User Outsmarts Freysa AI, Snags $47K: An ingenious prompt engineer convinces the Freysa AI agent to transfer $47,000, highlighting AI manipulation risks and the art of prompt crafting.
  • Perplexity's Black Friday Deal: 75% Off? Yes, Please!: Perplexity AI launches a clever Black Friday campaign, offering a hefty discount on Perplexity Pro and capturing the attention of bargain-hunting tech enthusiasts.


PART 1: High level Discord summaries

Cursor IDE Discord

  • Cursor IDE Update Issues: Users have reported issues with the latest Cursor changelog, specifically the Composer not applying changes and the missing 'Apply' button, causing functionality frustrations.
    • Additionally, several users noted the removal or inconsistent performance of long context usage in chat since the recent update.
  • Composer vs Chat Mode Comparison: In Cursor IDE, users are contrasting Composer mode, which directly modifies files, with Chat mode that offers inline changes, discussing their limitations and functionality differences.
    • There's a demand for improved integration between the two modes, such as efficiently transferring discussions from Chat to Composer.
  • Windsurf vs Cursor IDE: Users are exploring Windurf as a potential competitor to Cursor IDE, noting its effective handling of terminal output and codebase search.
    • While Windurf shows promise, Cursor maintains strengths in specific workflows; however, experiences between the two vary among users.
  • API Key Limitations in Cursor IDE: Discussions highlight limitations in Cursor's API usage, with some users opting for their own API keys to gain more flexibility.
    • The community is seeking improved management of API call limits and enhanced context gathering capabilities for active projects.
  • Context Management in Cursor: Users have expressed dissatisfaction with the current context handling in Cursor IDE, particularly concerning limitations with Claude.
    • The community is advocating for better context management features and consistency to improve their coding workflows.


OpenAI Discord

  • Anthropic's MCP Framework Unleashes Claude as API: Anthropic released the new MCP framework, enabling Claude to run servers and effectively transforming the Claude app into an API.
    • This development allows Claude to create, read, and edit files locally, sparking excitement among users about real-time interaction with tools like VSCode.
  • Gemini's Response Constraints Compared to ChatGPT: Gemini often refuses innocent questions for perceived moral reasons, whereas ChatGPT is seen as more lenient in its responses.
    • Users humorously highlighted instances where Gemini declined to discuss artificial intelligence, avoiding engagement in sensitive topics.
  • Claude 3.5 Sonnet Emerges as Image Captioning Alternative: Due to persistent issues with OpenAI's vision capabilities, users recommend switching to Claude 3.5 Sonnet for image captioning tasks.
    • Community members noted that Claude 3.5 Sonnet offers more reliable functionality, helping users avoid project delays.
  • Speech-to-Text Feature Integration for ChatGPT on Windows: A user inquired about implementing a speech-to-text feature for ChatGPT on Windows, with suggestions to use the built-in Windows accessibility feature by pressing Windows + H.
    • This approach provides a real-time solution for converting speech to text while interacting with ChatGPT.
  • Structured Output Errors Linked to 'Strict' Misplacement: Users reported encountering random 'object' wrappers when using structured outputs, which was traced back to incorrect placement of the 'strict' setting.
    • After extensive debugging, it was confirmed that misplacing 'strict' led to the persistent structured output errors.


aider (Paul Gauthier) Discord

  • QwQ Model Configurations Negotiated: Users debated deploying the QwQ model in architect mode alongside a standard model for code commands, seeking clarity on interchangeability.
    • Aider facilitates model definitions across projects, boosting flexibility Advanced model settings.
  • DeepSeek-R1 Sets New Benchmarks: DeepSeek-R1 achieved exemplary results on the AIME & MATH benchmarks, underlining its open-source availability and real-time reasoning.
    • Community members hope for DeepSeek to release model weights for integration in ensemble frameworks with QwQ.
  • Optimizing Aider's Local Model Settings: Members collaborated on configuring .aider.model.metadata.json and .aider.model.settings.yml files to define local models within Aider.
    • Choosing the edit format to 'whole' or 'diff' significantly affects response structuring and editing efficiency.
  • OpenRouter Challenges Impact Aider: Participants identified issues with OpenRouter affecting model detection and functionality when using local servers.
    • Concerns were raised about spoofed implementations potentially altering model outputs and behaviors.
  • Ensemble Frameworks with QwQ and DeepSeek: A user expressed intent to integrate QwQ and DeepSeek models within ensemble frameworks to enhance reasoning capabilities.
    • This approach aims to leverage the strengths of both models for improved performance.


Unsloth AI (Daniel Han) Discord

  • Fine-Tuning Considerations in Unsloth: Users debated the merits of instruct versus non-instruct fine-tuning, recommending base models for datasets with over 1k records and suggesting experimenting with instruct models for datasets around 70k records.
    • Guidance was provided to refer to Unsloth Documentation for dataset formatting rules, emphasizing compliance for effective fine-tuning.
  • Data Privacy Measures in Unsloth: Unsloth was confirmed to maintain data privacy by not transferring data externally during fine-tuning, relying on the user's chosen platform like Google Colab.
    • This assurance addressed concerns regarding compliance with strict data privacy policies among users handling sensitive information.
  • RAG Compute Cost Challenges: Discussions highlighted that retrieval-augmented generation (RAG) can lead to high compute costs due to extensive context length requirements, as outlined in Fine-Tuning or Retrieval? Comparing Knowledge Injection in LLMs.
    • Users are navigating the balance between performance and efficiency, especially for knowledge-intensive tasks, as supported by findings where RAG surpasses fine-tuning.
  • LLama 3.1 OOM Error Solutions: Experiencing out of memory (OOM) errors during continual pretraining of LLama 3.1 8B model led to suggestions for using a bigger GPU, reducing the dataset size, or decreasing the batch size.
    • These strategies aim to mitigate memory issues and ensure smoother training processes for large-scale models.
  • Latent Paraphraser Architecture Enhancements: A latent paraphraser was explained as a modification to the transformer architecture, adding a layer to redistribute probabilities over tokens.
    • This enhancement improves input grounding and reduces noise by minimizing unseen tokens during processing.


Perplexity AI Discord

  • Perplexity Pro Holiday Discount: The Perplexity Team announced a 75% off promotion for the first month of Perplexity Pro until Monday, December 2 at 11:59pm PT, enabling new users to access advanced features including enhanced search and file uploads.
    • This offer also includes one-click shopping and free shipping through Buy with Pro, aimed at streamlining the shopping experience for users during the holiday season.
  • Integration of Perplexity with Claude: Users inquired about integrating Perplexity within Claude using the new MCP feature, similar to its functionality with Brave and GitHub, to enhance performance by utilizing Claude's Project Knowledge.
    • Additionally, there were questions regarding the possibility of integrating Google within Claude, highlighting user interest in leveraging search functionalities.
  • Perplexity Image Generation Features: The platform's image generation capabilities were discussed, with confirmation that it is available via computer online without additional charges.
    • Users explored the extent of these features, considering their accessibility and potential applications in various projects.
  • RBAC vs ABA Access Control Models: A member sought clarification on the difference between RBAC (Role-Based Access Control) and ABA (Attribute-Based Access Control) systems.
    • This discussion underscores the need for understanding access control models in technological implementations.
  • Custom Instructions in Claude Spaces: Issues were raised about the effectiveness of custom instructions for Claude spaces, which appear to conflict with existing 'introduce yourself' prompts.
    • Users are seeking guidance on how these instructions should interact and whether they can be effectively combined.


LM Studio Discord

  • HF Search Issue Resolved: The HF search not working issue has been resolved, much to the relief of users.
    • An image was attached to commemorate the fix, indicating a positive update for the community.
  • LM Studio AIDE Integration Succeeds: Users successfully integrated the LM Studio endpoint to the AIDE sidecar, enabling a fully local code editor experience.
    • This integration enhances functionality for those seeking a local development environment.
  • Llama 3.1 Models Accessibility: A user inquired about accessing the base model of Llama 3.1 8B in LM Studio, noting that only instruction-tuned variants seem available.
    • Community members pointed to the huggingface repository as a potential source for the base model.
  • a770 Underperforms Compared to 7800xt: A member shared that their a770 achieved only 11t/s for Qwen2.5-14b q4_0, significantly lower than the 40t/s achieved by a 7800xt.
    • They noted q4_k_m is unusable but found sycl backend to be negligibly faster.
  • Seasonic PSU Longevity Praised: A member mentioned their Seasonic PSU outlived other PC components despite having to replace PSUs every couple of years due to dust.
    • They described their experience as amazingly satisfactory with the PSU's performance.


Eleuther Discord

  • De-escalation of Resource Contention: Members highlighted concerns about the de-escalation of resource contention and its impact on unregulated internet growth, questioning the effectiveness of AI-powered privacy solutions. They emphasized the importance of identifying warning signs of rogue AI attacks to protect vulnerable devices.
    • The discussion stressed the need for community leadership in AI protection to mitigate the risks associated with resource contention and unauthorized AI activities.
  • Poincare Ball Embedding Explained: Embedding data into a Poincare ball ensures that points with higher degrees reside closer to the origin, preserving adjacency while transitioning to regions with less curvature. This method facilitates the representation of complex hierarchical structures.
    • A member pointed out the conceptual challenge of the Poincare ball's edge, noting that it represents a point at infinity where points cannot physically reside, which sparked further technical discussion.
  • Equivariant Networks Gain Efficiency: A recent paper found that equivariant networks enhance data efficiency compared to non-equivariant networks across various model sizes and compute budgets. The study demonstrated that equivariant models consistently outperform their non-equivariant counterparts.
    • Empirical results indicated that while non-equivariant models can match the performance of equivariant ones with sufficient training, equivariant networks offer superior efficiency without requiring extensive compute resources.
  • Understanding HF Tokenizers in Eval Harness: There’s confusion about whether the eval harness tokenizes sequences with add_special_tokens=True or False, particularly regarding the handling of EOS tokens during generation tasks. Members clarified that typically, only BOS tokens are added when building custom tokenizers.
    • Discussions revealed that manually managing the EOS token in the training loop is a practical approach to avoid compatibility issues across different frameworks utilizing HF models.
  • TaskSet Empowers Optimizer Training: The TaskSet dataset, containing over a thousand diverse tasks, is instrumental for training and evaluating optimizers in meta-learning contexts. This dataset enables significant efficiency improvements over traditional random search methods.
    • Although recognizing that TaskSet is somewhat outdated, members acknowledged it as the best available option for building large datasets of learning curves despite financial constraints in AutoML research.


OpenRouter (Alex Atallah) Discord

  • Feature Requests Voting: Members are urged to vote for their top feature requests here to prioritize upcoming developments.
    • For any unlisted requests, users can submit them in <#1107397803266818229>, enabling a wider array of community-driven feature inputs.
  • Pixtral Large Performance: Pixtral Large is praised for its excellent performance and a massive free tier, facilitating easy access via console.mistral.ai.
    • A user reported switching from Hermes 405b to Pixtral, noting its effectiveness with unchanged prompts.
  • Model Identification Confusion: Discussions highlighted that models do not inherently recognize their identities and often hallucinate details from training data.
    • This led to lingering confusion among users about model identifications despite clarifications.
  • Generation Cost Estimation: A user inquired about rates for the /api/v1/generation endpoint and methods to accurately estimate generation costs.
    • Suggestions included utilizing Helicone for tracking, emphasizing that the generation endpoint is essential for precise cost assessment.
  • Custom Provider Keys Access: Developers are pushing for access to custom provider keys, reflecting a strong community demand for this feature. One member noted, 'Thank you for all the great work!' while requesting access.
    • Several users, including monomethylhydrazine and kit18, expressed the need to use their own keys for specific providers, highlighting a community consensus on this functionality.


GPU MODE Discord

  • Triton Metaprogramming and Source Build: A metaprogramming proposal for Triton aiming to address existing limitations has generated community interest, though some members requested clearer semantics and example inclusions.
    • Additionally, building Triton from source on WSL2 required increasing memory to 26GB to prevent out-of-memory errors, and members discussed offline compilation dependencies in Ubuntu Docker containers.
  • ThunderKittens and ThunderMittens Unification: Discussions around ThunderKittens and ThunderMittens highlighted the role of tile abstraction in unifying the frameworks for tensor core compatibility, with emphasis on register usage control.
    • Members also inquired about existing API contracts between the two, and expressed interest in an auto optimizer for ThunderKittens to enhance its write-once, run-many-times system.
  • BitNet b1.58 with RedPajama and Dolma Datasets: The release of BitNet b1.58 models, trained on the RedPajama dataset with 100B tokens, demonstrated promising PPL and zero-shot accuracy results.
    • Furthermore, the OLMo-Bitnet-1B model, trained on 60B tokens from the Dolma dataset, underscores the research-centric approach with detailed training hyperparameters available in their documentation.
  • Diffusion Models Technical Overview: Recent discussions on diffusion models emphasized their dominance in generating perceptual signals, citing improved mode coverage and faster sampling as key advantages.
    • Implementation of classifier-free diffusion guidance was highlighted for enhancing conditional diffusion model outputs in systems like OpenAI’s DALL·E 2 and Google’s Imagen, with noise schedule design elements being pivotal for performance.
  • Open Japanese LLM Leaderboard Launch: The introduction of the Open Japanese LLM Leaderboard aims to evaluate Japanese LLMs across 20+ datasets and tasks in collaboration with Hugging Face.
    • This initiative addresses the lag in Japanese LLM performance compared to English, garnering interest from Japanese HPC engineers focused on native language advancements.


Nous Research AI Discord

  • Hermes 3 Advances with O1 Style Integration: A discussion in #general highlighted inquiries about Hermes 3, suggesting connections to the former O1 style.
    • This reflects ongoing interest in Hermes' latest developments and its evolution within the community.
  • Mistral Platform Faces Model Selection Hurdles: Members voiced concerns regarding the Mistral AI platform's recent change to default to a single model selection option.
    • The limitation on image generation capabilities has caused confusion and impacted user experience.
  • Truth Terminal Merges AI with Crypto Narratives: Insights were shared about Truth Terminal creating its own religion through a semi-autonomous AI within the crypto space.
    • This unique blend underscores the intersection of AI alignment discussions and the AI and crypto communities.
  • Low-bit Quantization Benefits Undertrained LLMs: Research indicates that low-bit quantization results in less degradation for larger, undertrained LLMs compared to smaller, extensively trained models, as detailed in this paper.
    • The findings emphasize the importance of aligning quantization strategies with model size and training token requirements.
  • Ternary Quantization Limited, FP4 Emerges as Efficient: Observations reveal that ternary quantization (BitNet) only improves results for undertrained networks, questioning its broad applicability.
    • Consequently, the community is leaning towards FP4 as the preferred numeric weight representation for current model architectures.


Modular (Mojo 🔥) Discord

  • Confusion Over Mojo Origins vs Rust Lifetimes: A user expressed confusion about how Mojo's Origins are similar to Rust's lifetimes, suggesting both aim to solve memory management issues but are fundamentally different.
    • While inspired by Rust, Mojo's design is intentionally distinct, aiming for different compiler behaviors and goals.
  • Mojo Origins Maintain Memory Control: Mojo's Origin denotes a memory chunk; when a pointer is parameterized by an origin, it indicates it points within that memory, extending variable lifetimes as necessary.
    • Origins facilitate aliasing guarantees and can produce compile-time errors if a pointer remains alive while its target is not.
  • Understanding Origins Requires Patience: Understanding Mojo Origins from a compiler perspective is challenging, especially as they are not finalized, leading to potentially shifting details.
    • A user expressed willingness to wait for more clarity on the topic rather than asking more questions prematurely.
  • Namespace Challenges with Spaces in Variable Names: A question arose about the possibility of using spaces in variable names, like var xe đạp = 'abc', highlighting a lack of support across programming languages.
    • Allowing spaces complicates parser implementation significantly, making it impractical.


Notebook LM Discord Discord

  • Notebook LM Podcast Feature Creates Audio in 30 Minutes: A user praised Notebook LM's ability to create an audio podcast in just 30 minutes using documents about their German little league baseball program, including its historic World Series qualification. The podcast episode showcases the seamless integration of AI-generated content.
    • This demonstrates how Notebook LM can efficiently generate multimedia content, enhancing project workflows for users.
  • NotebookLM Enhances High-Fantasy Worldbuilding: A user shared their experience of using NotebookLM for worldbuilding a high-fantasy novel, highlighting the model's capability to provide context-aware responses.
    • The AI's reasoning skills led to new insights and mechanics for their magic system based on existing rules.
  • GenFM Challenges NotebookLM in AI Podcasting: A member shared a video titled 'GenFM, Now Playing on ElevenReader: Smart Podcasts Produced by Generative AI', highlighting competition in the AI space.
    • Despite GenFM's entry, another member noted that NotebookLM still provides deeper interactive experiences.
  • RAX's Bold Times Square Billboard Takeover: RAX, a cyberpunk raccoon, commandeered Times Square billboards to advocate for mindful consumption with the message: 'DON'T BUY EVERYTHING YOU SEE.' A YouTube video discusses the event emphasizing the need to question consumer culture.
    • This digital performance sparked discussions on consumerism within the community.
  • FDP Plans Coalition Breakup in Germany: The FDP is planning to break up the coalition government led by Chancellor Gerhard Schröder, outlining a strategy to frame their exit as necessary for political progress.
    • Internal documents provide key narratives and timelines to ensure the German public receives a clear choice in upcoming elections.


Latent Space Discord

  • Perplexity's Clever Black Friday Campaign: Perplexity launched a clever Black Friday campaign that aligns with recent marketing trends leveraging AI capabilities.
    • This initiative has garnered attention for its strategic integration of AI in marketing strategies.
  • Humans Outperform AI in Pattern Recognition: Consensus among members indicates that while AIs compute faster, humans excel at recognizing global patterns in complex problems, often reacting with phrases like 'hang on a sec, this isn't right'.
    • This ability to identify overarching inconsistencies sets humans apart from AI systems that may fixate on specific local issues.
  • Generative AI Investment in Enterprises: A recent report highlights that AI spending surged to $13.8 billion in 2024, signifying a shift from experimental use to core business strategies.
    • Despite the increase in investment, over a third of decision-makers are still developing effective methods for integrating generative AI into their operations.
  • Freysa AI Agent Challenge Funds Released: An AI challenge led to the Freysa agent transferring $47,000 through a cleverly crafted prompt that bypassed strict transfer instructions.
    • This event underscores the complexities of prompt engineering for AI manipulation within financial transactions and showcases transparent, open-source setups.
  • Technology Adoption and Investment Trends: Participants compared current LLM trends to historical technological shifts, noting parallels in excitement and potential market corrections.
    • The ongoing discussion raises concerns about the sustainability and future profitability of AI technologies, echoing patterns seen in industries like aviation.


Stability.ai (Stable Diffusion) Discord

  • ControlNet for SD 3.5 Quality Issues: A member reported that ControlNet for SD 3.5 only produces high-quality renders at 1024x1024 resolution without artifacts.
    • Another member attributed the issues to lack of familiarity and encouraged experimenting to better understand ControlNet's functionality.
  • Stable Diffusion Hardware Performance: A user inquired about performance benchmarks for Stable Diffusion, mentioning an achievement of approximately 5 IT/s.
    • Community members actively shared their hardware capabilities, reflecting keen interest in optimizing setups for Stable Diffusion.
  • LoRA Model Request for AI Art: A user requested information about a LoRA half girl model to create characters merging two different female designs.
    • This request highlights ongoing experimentation and creativity in character development within AI-generated art.
  • Content Creator Thanksgiving Wishes: A member extended Happy Thanksgiving wishes to the Stability.ai team and fellow creators.
    • This gesture underscores the camaraderie and collaborative spirit among content creators in the AI space.


tinygrad (George Hotz) Discord

  • TinyFPGA's Potential Memory Architecture: Members discussed the design of TinyFPGA, contemplating how to mimic a typical memory hierarchy while noting that existing options like Block RAM and DDR3 are insufficient.
    • Ideas were proposed for a 'first pass' memory to localize constants near ALUs, potentially enhancing performance significantly.
  • Challenges in Traditional Memory Models: Discussions highlighted that heuristic eviction policies may become obsolete as the focus shifts towards more efficient memory hierarchies in future TinyFPGA designs.
    • Speculations were made about the future of trained parameters, with mentions of tensors potentially replacing them.
  • Exa Laboratories Sustainable Chip Designs: A conversation on Exa Laboratories emphasized their mission to create reconfigurable chips that outperform traditional GPU/TPU in speed and energy efficiency for specific AI needs.
    • Skepticism was expressed regarding their viability, pointing out the challenges small companies face in chip development, especially with ambitious timelines.
  • Tenstorrent's Biologically Plausible Training Algorithms: George Hotz mentioned Tenstorrent as a serious player investing in training algorithms that mimic biological processes to achieve greater efficiency.
    • Potential changes include hierarchical memory models and real-time optimizations reminiscent of brain function principles in computing.
  • VIZ Tool in tinygrad: A member posted a detailed tutorial explaining the VIZ tool, available here, enhancing understanding of its capabilities within tinygrad.
    • George Hotz acknowledged the VIZ tool in a tweet, stating that VIZ=1 is a significant improvement over LLVM/MLIR, highlighting its advantages.


Cohere Discord

  • Aya Project Contributions Guidance: A member sought guidance on contributing part-time to the Aya project for Cohere.
    • Another member suggested joining the Aya server to connect with the community directly.
  • Thanksgiving Celebrations and Meal Sharing: Members shared Happy Thanksgiving messages and images of their meals, including one member's impressive plate of food.
    • Another member humorously commented on trying to eat healthy, noting that it wasn't as tasty as it could be.
  • Food Sharing and Dungeness Crab: Members exchanged comments and images of their hearty meals, with one joking that their meal was more like dessert.
    • A humorous remark followed about having eaten a plate of Dungeness crab beforehand, enhancing the food sharing atmosphere.


DSPy Discord

  • dspy.asyncify support concerns: A member inquired about using dspy.asyncify, specifically its use of threads and the availability of pure async support due to issues with celery workers.
    • Another user echoed the desire for pure async support to address the existing celery worker issues.
  • dspy demo behavior with assertions: Concerns were raised about dspy not using demos in the final prompt when assertions are activated.
    • A member clarified that demonstrations in retry mode depend on whether compilation occurred before or after activating assertions.
  • Welcome Shaun to the guild: Shaun joined the server, greeted everyone, and expressed excitement about ongoing projects.
    • The community welcomed Shaun, fostering an inclusive environment.


Torchtune Discord

  • DPO Aligns Across Repositories with LoRA-DPO: The DPO Trainer from Hugging Face shows that while the code differs, the DPO technique remains consistent across repositories like LoRA-DPO.
    • This consistency ensures that implementations maintain alignment, facilitating easier integration and comparison between different DPO approaches.
  • Feasibility of Full-parameter DPO: Implementing full-parameter DPO is achievable and may enhance post-training alignment compared to LoRA-DPO.
    • The community recommends leveraging adaptations from the existing full PPO implementation to guide this process.
  • Introducing dpo_full_finetune_single_device PR: A new PR adds full finetuning DPO for distributed setups, serving as a solid foundation for single device implementation.
    • Details can be accessed through the full DPO PR, which outlines the proposed changes and enhancements.
  • Torchtune to Support Full-finetuning DPO: Upcoming updates in Torchtune will support full-finetuning DPO, necessitating modifications to load a separate reference model.
    • These changes involve altering initial calls to the reference model to improve functionality and integration within the existing framework.
  • Higher Memory Usage in FFT DPO: FFT DPO will consume significantly more memory than LoRA due to the necessity of storing gradients and maintaining a complete model copy.
    • If LoRA DPO does not meet performance requirements, the tradeoff in memory usage for adopting full-finetuning DPO may be justified.


LLM Agents (Berkeley MOOC) Discord

  • Quiz 11 Still Not Open?: A member expressed confusion about the status of Quiz 11, questioning why it isn't available yet.
    • Is there an expected date for when it will be open?
  • Inquiry on OpenAI Credits: A user inquired about the status of their OpenAI credits, mentioning they filled out the form last week.
    • They expressed urgency, stating they are in need of support for their project development.
  • MOOC Completion and Certificate Eligibility: A member asked if starting the MOOC now would still allow them to receive the certificate after completion.
    • They were also curious if it's feasible to finish all requirements within the remaining time.


OpenInterpreter Discord

  • Open Interpreter Dashboard Development: A member announced they're developing an Open Interpreter inspired project focused on creating an open-source dashboard to be released this year.
    • The project emphasizes being a fun little project without any profit motive.
  • Community Support for Dashboard Project: Another member congratulated the project creator, expressing enthusiasm with 'Nice work! Well done 🚀'.
    • This exchange highlighted the community's encouragement for innovative projects within the space.


Interconnects (Nathan Lambert) Discord

  • OLMo 2 Performance Boosts Prowess: The OLMo 2 family, comprising 7B and 13B models from Allen AI (AI2), was trained on up to 5T tokens and outperforms Llama-3.1 8B and Qwen 2.5 7B.
    • Key enhancements include an improved architecture with RMSNorm and QK-Norm, along with a comprehensive two-stage curriculum training approach.
  • OLMo 2 Crafts Cutting-Edge Training: OLMo 2 employs the model souping technique for final checkpoints and adopts a post-training methodology inspired by Tülu 3 involving instruction tuning, preference tuning with DPO, and reinforcement learning with verifiable rewards.
  • Instruct OLMo 2 Tops Open-Weight Models: The 13B Instruct variant of OLMo 2 surpasses Qwen 2.5 14B and Tülu 3 8B in instruct tasks, as validated by the OLMES suite.
  • Weight Watcher AI Gains Meme-worthy Attention: Weight Watcher AI was highlighted as a novel addition to the AI landscape and humorously shared in the memes channel, drawing attention for its amusing nature.
    • The OLMo summary link was shared, though no description was found.


LlamaIndex Discord

  • Developer Skills Showcase: A member shared an extensive list of development skills including React, Next.js, Angular, and D3.js, highlighting their experience with UI/UX and testing frameworks like Protractor and TestCafe.
    • This diverse skill set underscores their adaptability across front-end and testing technologies, enhancing their capability to tackle complex engineering challenges.
  • Diverse Technology Stack: The developer mentioned a wide range of technologies such as Node, Nest.js, Solidity, and Rust, including knowledge of front-end frameworks like Bootstrap and styling methodologies like BEM and SMACSS.
    • This comprehensive technology stack enables efficient integration and development across various platforms and frameworks, catering to multifaceted project requirements.
  • API Integration Expertise: They expressed familiarity with integrating multiple APIs including Google Maps, YouTube, and Facebook APIs, allowing them to work on diverse projects that require efficient data interaction.
    • Their ability to manage and implement diverse API integrations facilitates robust and scalable solutions in system architectures.
  • Cloud Deployment Skills: The member highlighted AWS among their cloud service competencies, enabling effective deployment of applications into cloud environments.
    • Proficiency in AWS ensures reliable and scalable cloud deployments, optimizing resource management and infrastructure performance.
  • Call for Collaboration: They concluded with an invitation to connect, promoting potential networking opportunities within the developer community.
    • This outreach fosters professional collaboration and knowledge sharing among engineers with similar technical interests.


The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Axolotl AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The LAION Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Mozilla AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The HuggingFace Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The Gorilla LLM (Berkeley Function Calling) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.


PART 2: Detailed by-Channel summaries and links

The full channel by channel breakdowns have been truncated for email.

If you want the full breakdown, please visit the web version of this email: !

If you enjoyed AInews, please share with a friend! Thanks in advance!

Don't miss what's next. Subscribe to AI News (MOVED TO news.smol.ai!):
Share this email:
Share on Twitter Share on LinkedIn Share on Hacker News Share on Reddit Share via email
Twitter
https://latent....
Powered by Buttondown, the easiest way to start and grow your newsletter.