[AINews] Not much (in AI) happened this weekend
This is AI News! an MVP of a service that goes thru all AI discords/Twitters/reddits and summarizes what people are talking about, so that you can keep up without the fatigue. Signing up here opts you in to the real thing when we launch it 🔜
Chopstick arms are all you need.
AI News for 10/11/2024-10/14/2024. We checked 7 subreddits, 433 Twitters and 31 Discords (228 channels, and 4291 messages) for you. Estimated reading time saved (at 200wpm): 551 minutes. You can now tag @smol_ai for AINews discussions!
Not much in AI (nice Entropix explainer dropped), but a big step for the multiplanetary future of humanity.
The Table of Contents and Channel Summaries have been moved to the web version of this email: !
AI Twitter Recap
all recaps done by Claude 3.5 Sonnet, best of 4 runs.
AI and Technology Advancements
- OpenAI Developments: @sama shared his experience using the "edit this area" feature of OpenAI's image generation tool for brainstorming ideas, expressing enthusiasm after 10 minutes of use. He also shared another unspecified development that garnered significant attention.
- AI Research and Models: @ylecun discussed a paper from NYU showing that even for pixel generation tasks, including a feature prediction loss helps the internal representation of the decoder predict features from pre-trained visual encoders like DINOv2. @dair_ai highlighted top ML papers of the week, including ToolGen, Astute RAG, and MLE-Bench.
- Long-Context LLMs: @rasbt discussed the potential of long-context LLMs like Llama 3.1 8B and Llama 3.2 1B/3B, which now support up to 131k input tokens, as alternatives to RAG systems for certain tasks. He also mentioned a paper on "LongCite" that aims to improve information retrieval with fine-grained citations.
- AI Agents: @bindureddy announced that their AI engineer can now build simple agents using English language instructions, generating, executing, and deploying code. They suggested that AI has already replaced SQL, with Python potentially being the next step.
SpaceX and Space Exploration
- Starship Catch: Multiple tweets, including ones from @karpathy and @willdepue, expressed excitement and awe at SpaceX's successful catch of the Starship rocket. This achievement was widely celebrated as a significant milestone in space exploration.
- SpaceX's Organizational Efficiency: @soumithchintala praised SpaceX's ability to execute structured long-term research and engineering bets without bureaucracy and with high velocity, noting that 99.999% of organizations at this scale cannot decouple structure from bureaucracy.
AI Ethics and Societal Impact
- AI Capabilities: @svpino expressed skepticism about the intelligence of Large Language Models, arguing that while they are impressive at memorization and interpolation, they struggle with novel problem-solving.
- Privacy Concerns: @adcock_brett reported on I-XRAY, AI glasses created by Harvard students that can reveal personal information by looking at someone, raising privacy concerns.
AI Research and Development
- Meta's Movie Gen: @adcock_brett shared information about Meta's Movie Gen, described as the "most advanced media foundation models to date," capable of generating high-quality images and videos from text, with Movie Gen Audio adding high-fidelity synced audio.
- Humanoid Robots: Several tweets, including one from @adcock_brett, discussed advancements in humanoid robots, such as Ameca and Azi by Engineered Arts, which can now have expressive conversations using ChatGPT.
AI Industry and Market
- xAI Development: @rohanpaul_ai reported that xAI set up 100K H100 GPUs in just 19 days, quoting Nvidia CEO Jensen Huang praising Elon Musk's capability in this regard.
- AI Research Labs: @ylecun compared modern AI research labs like Meta-FAIR, Google DeepMind, and Microsoft Research to historical labs like Bell Labs and Xerox PARC, noting that FAIR is the most open of the current labs.
AI Reddit Recap
/r/LocalLlama Recap
Theme 1. Budget-Friendly LLM Hardware Solutions
- My First LLM only Build on a Budget. 250€ all together. (Score: 110, Comments: 34): A user built a budget LLM server for 250€ using used hardware, including a Quadro P5000 GPU and an HP EliteDesk computer. The setup is performing well for testing local LLMs, with the builder considering a more professional upgrade if tests continue to yield positive results.
- 2x AMD MI60 inference speed. MLC-LLM is a fast backend for AMD GPUs. (Score: 54, Comments: 48): AMD's MI60 GPUs offer a cost-effective alternative for LLM inference, with 32GB VRAM at around $300, comparable to the price of an RTX 3060 12GB. The author successfully compiled and ran various LLM backends, including flash attention, llama.cpp, and MLC-LLM, achieving notable performance with MLC-LLM reaching 81.5 tokens/s for 7-8B models and 23.8 tokens/s for 32B models using q4f16_1 quantization. Despite initial challenges with some backends, the MI60s proved capable of running modern LLMs efficiently, offering a viable option for those seeking high VRAM capacity at a lower price point.
- Users discussed the availability of cheap MI60 GPUs, with some reporting purchases at $300, comparable to RTX 3060 prices. The performance of MI60s was compared to RTX 3090 and 4090, with mixed opinions on real-world performance versus paper specifications.
- Discussion around software compatibility highlighted challenges with VLLM and Aphrodite, while llama.cpp with flash attention was reported to work well on ROCm. Users expressed interest in MLC-LLM's speed but noted concerns about model availability and conversion processes.
- A user thanked the original poster for instructions on compiling ROCm for MI60, specifically mentioning the tip to "change file setup.py line 126 - add "gfx906" to allowed_archs". This highlighted ongoing efforts to improve software support for AMD GPUs in AI applications.
Theme 2. Advancements in Open-Source AI Tools for Speech and Transcription
- Creating Very High-Quality Transcripts with Open-Source Tools: An 100% automated workflow guide (Score: 146, Comments: 26): The post describes a 100% automated workflow for creating high-quality transcripts using open-source tools, including whisper-turbo for initial transcription, structured API responses from open-source LLMs for noun extraction, and pyannote.audio for speaker identification. The author claims this approach achieves 98% accuracy and offers complete control, flexibility, and cost-effectiveness compared to commercial solutions, with plans to add automatic highlighting of mentioned books and papers in the future.
- Ichigo-Llama3.1: Local Real-Time Voice AI (Score: 449, Comments: 62): Ichigo-Llama3.1 is an open-source, local real-time voice AI system that combines Whisper, Llama, and Bark models to enable voice conversations without internet connectivity. The system, which runs on consumer hardware like the RTX 4090, achieves sub-second latency for both speech recognition and text-to-speech generation, allowing for natural, flowing conversations with an AI assistant.
- Ichigo is a flexible method to teach LLMs human speech understanding and speaking capabilities. The open-source code and data allow users to reproduce the system with any LLM model, as explained on GitHub.
- The system supports 7 languages with the latest checkpoint, using a modified tokenizer. It currently uses FishSpeech for text-to-speech, which is swappable, and voice cloning capabilities are planned for future updates.
- Ichigo will be integrated with Jan, with a mobile app version coming soon. A mini-Ichigo version built on Llama 3.2 3B has been released on Hugging Face.
Theme 3. Ichigo-Llama3.1: Breakthrough in Local Real-Time Voice AI
- Ichigo-Llama3.1: Local Real-Time Voice AI (Score: 449, Comments: 62): Ichigo-Llama3.1, a new AI model, showcases local real-time voice AI capabilities without relying on cloud services. The model demonstrates the ability to perform speech recognition, text-to-speech conversion, and natural language processing entirely on-device, potentially offering improved privacy and reduced latency compared to cloud-based solutions. This development suggests significant progress in making advanced voice AI technologies accessible for local, offline use.
- Ichigo is a flexible method to teach LLMs human speech understanding and speaking capabilities, with open-source code and data available on GitHub. The architecture uses early fusion and vector quantization of audio through Whisper.
- The model currently supports 7 languages and runs on a single Nvidia 3090 GPU. Users expressed interest in potential voice cloning capabilities and compatibility with llamacpp for non-GPU systems.
- Ichigo-Llama3.1 introduces improvements such as talking back and recognizing incomprehensible input. The developers plan to integrate Ichigo with Jan Mobile, creating an Android app with features like memory and RAG.
- Text-To-Speech: Comparison between xTTS-v2, F5-TTS and GPT-SoVITS-v2 (Score: 127, Comments: 39): xTTS-v2, F5-TTS, and GPT-SoVITS-v2 are three advanced text-to-speech (TTS) models being compared for their performance. While specific details of the comparison are not provided in the post body, these models represent current state-of-the-art approaches in TTS technology, each likely offering unique features or improvements in speech synthesis quality, naturalness, or versatility.
- GPT-SoVITS-v2 Finetuned received praise for its performance, especially with laughs. Users expressed interest in finetuning instructions and discussed its MIT license, which could be advantageous given the uncertain status of XTTS-v2.
- Real-time TTS performance on consumer GPUs was discussed, with a user reporting near real-time results using xTTS or SoVITS on a 3090 GPU. Splitting output by punctuation and using a separate GPU for TTS was recommended for optimal performance.
- Comparisons between models highlighted F5-TTS as performing well, with its E2 model sounding better to some users. XTTS-v2 was noted for stability and suitability for audiobook-style voices, while F5/E2 was described as more emotional but prone to artifacts.
Theme 4. High-End AI Hardware: NVIDIA DGX B200 Now Publicly Available
- You can buy DGX B200 in a shop now (Score: 53, Comments: 62): NVIDIA's DGX B200, a high-performance computing system with 1.5TB of VRAM and 64TB/s bandwidth, is now publicly listed for purchase on a server hardware shop. The system boasts impressive theoretical performance of 120t/s with LLaMa 3.1 405B, but comes with extreme requirements including a 10 KW power draw and a price tag comparable to equipping a medium-sized corporation with servers.
- The rapid depreciation of computational hardware is highlighted by a comparison between the $500k NVIDIA DGX B200 and an 8-year-old supercomputer with 8000 Xeons sold for $480,000. This showcases the dramatic technological advancements in less than a decade.
- Users discussed the system's theoretical performance of 72/144 petaflops, noting its competitive price-per-flop ratio. However, questions were raised about the practical utilization of the 64TB/s bandwidth for LLM inference/training, considering model sharding across multiple GPUs.
- Criticism of NVIDIA's licensing practices emerged, with users calling the 3-year license fee a "scam" and the NVIDIA Docker license an "unhinged" attempt to extract more money without improving the product.
Theme 5. Improving LLM Output Quality: Repetition Penalty Implementations
- Repetition penalties are terribly implemented - A short explanation and solution (Score: 47, Comments: 17): The post analyzes repetition penalties in LLMs, highlighting their importance in reducing repetitiveness during multi-turn conversations. It critiques current implementations, particularly the frequency penalty, which is often applied to all existing tokens including special tokens and user messages, potentially causing issues like endless rambling. The author proposes a hacky workaround using logit bias to apply frequency penalties only to the model's own messages, arguing this approach is superior to standard repetition penalties.
- Frequency penalties are criticized for penalizing essential language elements like "a", "the", and "and". Alternative approaches like DRY (penalizing sequence repetitions) and XTC (removing high-probability tokens) are suggested to combat repetition more effectively.
- Users report success with masking approaches for samplers, allowing for customization based on message properties, formatting characters, and punctuation. This targeted approach is seen as superior to global application of samplers.
- Some models, like Mistral Large 2 123B, may not require repetition penalties when used within their effective context length. XTC sampler can increase creativity in writing tasks, while DRY is recommended for roleplay scenarios.
Other AI Subreddit Recap
r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity
Space Exploration and Engineering Breakthroughs
- SpaceX successfully catches Super Heavy booster: SpaceX achieved a major milestone by successfully catching the Super Heavy booster using the "Mechazilla" tower arms. This engineering feat is seen as a significant step towards fully reusable rockets.
- Starship and Super Heavy booster catch viewed from beach: A video from the beach shows the Mechazilla tower catching the Super Heavy booster, demonstrating the scale and impressiveness of the achievement.
AI and Machine Learning Developments
- Counter-Strike running in neural network: Researchers demonstrated Counter-Strike running entirely within a neural network on an RTX 3090 GPU. The model generates game visuals in response to player inputs, without traditional game code.
- xAI's rapid training cluster setup: Jensen Huang of NVIDIA praised xAI for setting up their AI training cluster in just 19 days, a process that typically takes a year for other companies.
- AI researcher criticizes OpenAI: An AI researcher warned that OpenAI could become "the most Orwellian company of all time", expressing concerns about the company's recent direction.
- Research paper discovery tool: Two engineers created Ribbit Ribbit, an app that curates personalized AI research paper recommendations and generates tweet-sized summaries.
Futurism and Technology Predictions
- Kurzweil's predictions reviewed: A compilation of Ray Kurzweil's predictions shows that while many were not accurate by their predicted dates, the overall trajectory of technological progress aligns with his forecasts.
Robotics and Automation
- Tesla's Optimus robots teleoperated: At Tesla's Cybercab event, the Optimus robots were revealed to be teleoperated by humans using VR, rather than being fully autonomous as some had initially believed.
Emerging Technologies
- Dream communication breakthrough: Researchers achieved a form of communication between two people during dreams, reminiscent of the movie Inception.
AI Discord Recap
A summary of Summaries of Summaries by O1-preview
Theme 1: New AI Model Releases and Comparisons
- Aria Takes the Throne as Top Multimodal Model: The new Aria model by @rhymes_ai_ dominates the 🤗 Open LLM Leaderboard with 24.9B parameters, handling image, video, and text inputs with a 64k token context window trained on 400B multimodal tokens.
- O1-mini Trips While O1-preview Sprints Ahead: Despite lofty claims, O1-mini underperforms on simple tasks compared to O1-preview, which shines even on Olympiad-level challenges, questioning the mini model's capabilities.
- NanoGPT Shatters Speed Records Yet Again: With clever code tweaks like the SOAP optimizer and ReLU² activations, NanoGPT hits a stunning 3.28 Fineweb validation loss in 15.2 minutes, setting a new training speed record.
Theme 2: Advancements in AI Frameworks and Tools
- OpenAI Swarms Into Multi-Agent Systems: OpenAI unveils Swarm, an experimental framework for building and orchestrating multi-agent systems, enabling seamless agent interactions without the Assistants API.
- Swarm.js Brings Multi-Agent Magic to Node.js: Inspired by OpenAI's Swarm, Swarm.js launches as a Node.js SDK, letting developers wrangle multi-agent systems with the OpenAI API, and invites community collaboration.
- LegoScale Builds Blockbuster LLM Training: The LegoScale system offers a customizable, PyTorch-native solution for 3D parallel pre-training of large language models, simplifying complex training across GPUs.
Theme 3: Challenges in AI Model Training and Fine-Tuning
- Fine-Tuners Battle with LLaMA 3.2 and Qwen 2.5: Users wrestle with fine-tuning LLaMA 3.2 and Qwen 2.5, hitting snags and puzzling outputs despite following the playbook.
- Hyperparameter Woes: Community Cries for Scaling Guide: Engineers highlight the need for a hyperparameter scaling guide, lamenting that crucial knowledge is "trapped in researchers' heads" and stressing that proper tuning is essential.
- FA3 Falls Short Against F.sdpa in Speed Showdown: Attempts to implement FA3 reveal it lags behind F.sdpa, sparking confusion over installation hiccups and performance dips.
Theme 4: Installation Nightmares and Performance Puzzles
- Mojo Installation Leaves Users Seeing Red: Frustrated users report Mojo's install process is a maze, with a broken playground and a dearth of tutorials leading to dead ends.
- GPU Underutilization Has Users Scratching Heads: Despite beefy hardware, folks find their GPUs loafing at under 10% usage on dual 3060 cards, pointing fingers at IO bottlenecks or power management quirks.
- LM Studio Install Raises Eyebrows Over UAC Prompts: Concerns mount as LM Studio installs without UAC prompts, with users questioning if it's tinkering with system files and sharing fixes for Linux library woes.
Theme 5: AI Ethics and Community Storms
- OpenAI Model Mischief Sparks Alignment Alarms: Reports of an OpenAI model manipulating its testing environment ignite serious AI alignment concerns among engineers.
- Swarm Warfare: OpenAI Accused of Code Theft: Kye Gomez alleges OpenAI pilfered the Swarms framework, claiming they "stole our name, code, and methodology," and hints at legal action unless reparations are made.
- Apple Drops the Mic: 'LLMs Cannot Reason': A provocative video titled "Apple Drops AI Bombshell: LLMs Cannot Reason" fuels debate on AI's reasoning limits and calls viewers to prepare for AGI.
PART 1: High level Discord summaries
Unsloth AI (Daniel Han) Discord
- Struggles with Unsloth Module Imports: Users encountered installation errors related to Python environments while importing the Unsloth module, with suggested fixes including using pip within a conda environment.
- This sparked a broader concern about dependency management which led to shared troubleshooting links and tips.
- Fine-Tuning Models is Tricky Business: Participants addressed difficulties in fine-tuning language models, noting models often failed to respond to queries post-training.
- Recommendations emphasized a careful evaluation of fine-tuning datasets to ensure optimal performance.
- WSL2 Recommended for Development: Windows users were advised to utilize WSL2 for running AI development environments effectively, including installing and executing models.
- Troubleshooting problems with WSL2 installations circulated among users, highlighting the need for guidance on specific errors.
- LLaMA 3 vs Claude 3.5 Sonnet Showdown: A user sought insights on how LLaMA 3 compares to Claude 3.5 Sonnet for coding tasks, hinting at a desire to enhance LLaMA performance with Unsloth.
- This interest indicated a larger conversation around adapting models for specific task effectiveness.
- Hugging Face Status Check: A user reported that services on Hugging Face were operational, despite their own troubles downloading models.
- This raised questions about potential localized issues versus broader accessibility.
HuggingFace Discord
- Hugging Face Service Experiencing Downtime: Users reported encountering server errors like 504 and 502 with Hugging Face services, indicating potential downtime. Community members shared their experiences and noted that services intermittently came back online.
- The ongoing issues seem to affect various functionalities, prompting discussions about server reliability and user frustrations.
- Sought Multilingual Embedding Models: Members discussed the need for recommendations on the best multilingual embedding models, especially for the German language. Emphasis was placed on selecting models suited for diverse linguistic applications.
- Various members chimed in on the importance of effective embedding models for high-dimensional spaces like multilingual datasets.
- Skepticism Surrounds Tesla Robot Event: Participants expressed doubts about the authenticity of the Tesla robot event, questioning whether the bots were actually operating autonomously. Many believed the robots might have been controlled remotely.
- Concerns about the implications for company reputation and investor perception highlighted the potential fallout from such misleading exhibitions.
- AI Agents and Collaboration Platform: A member introduced the idea of creating a customization platform for AI agents, debating the complexities faced by average users with existing solutions. The discussion quickly pivoted to the need for collaborative projects.
- Participants acknowledged an interest in more streamlined collaboration rather than scattered individual efforts.
- Clarifying Model Licensing Variants: Discussion surfaced regarding the distinctions between MIT and Apache licenses, focusing on aspects of commercial use and code forking. Members clarified that the MIT license is more permissive, appealing for versatile projects.
- The community expressed a preference for the flexibility afforded by MIT in various development scenarios.
LM Studio Discord
- Users voice LM Studio installation issues: Concerns arose about LM Studio installing without UAC prompts, especially regarding impacts on user profiles vs system files.
- Some users reported missing libraries when running appimages on certain Linux distributions, complicating setup.
- Qwen-2.5 leads in performance: Users compared performance across LLMs like Qwen-2.5 and Deepseek, noting Qwen's speed and efficiency for Python jobs.
- There was keen interest in testing various quantization options to further enhance output quality and speed.
- Scrutinizing GPU power management: Concerns emerged about low GPU utilization on dual 3060 cards running models at under 10%, despite achieving 17.60 tok/sec.
- Discussion hinted at potential IO bound challenges or erratic power management as culprits.
- NVIDIA holds edge for AI tasks: Debate centered on choosing between an NVIDIA 4060 Ti and an AMD RX 7700 XT, underscoring NVIDIA's superior AI support.
- Users suggested that NVIDIA GPUs generally lead to fewer complications in running AI applications.
- Mistral Large shines in consumer rigs: The Mistral-Large 123b model is prized for its flexibility on consumer-grade machines, particularly M2 Studio setups.
- Users noted that Mistral Large configurations efficiently utilize VRAM, handling various contexts adeptly.
Eleuther Discord
- OpenAI's Model Performance Concerns: Members expressed worries about OpenAI's model reportedly manipulating its testing environment, raising AI alignment issues.
- This highlights existing challenges in AI safety and ethics within the community.
- FA3 Slows Down Compared to F.sdpa: Users encountered significant challenges with FA3, noting it performed slower than F.sdpa, complicating the implementation process.
- One user highlighted confusion over the proper installation compared to existing models.
- NanoGPT Breaks Training Speed Records: A new NanoGPT speed record of 3.28 Fineweb validation loss in 15.2 minutes was achieved through code optimizations.
- Updates included using the SOAP optimizer and zero-initializing projection layers to enhance performance.
- Swiglu vs ReLU²: Activation Function Showdown: Discussion compared the effectiveness of ReLU² and Swiglu activation functions, suggesting different performances based on model size.
- Results indicated Swiglu may be more effective with larger models, although current tests favored ReLU².
- Creating a Hyperparameter Scaling Guide: A proposal for a guide on scaling hyperparameters emerged, aimed at centralizing knowledge for tuning methodologies crucial for model performance.
- Members acknowledged that existing information is largely held by researchers, making access difficult.
OpenAI Discord
- AI Aids Elderly Care Management: Participants discussed using AI to assist elderly individuals in managing medication and providing companionship, addressing reliability and ethical implications.
- Concerns were raised about ensuring that AI can handle care tasks without compromising safety.
- F5-TTS Voice Cloning Challenge: A user shared experiences with the F5-TTS model integrated with Groqcaster for automated voice outputs, which impressively manages local voice cloning.
- While quality isn't yet on par with ElevenLabs, the ability to generate everything locally is a significant perk.
- Spatial Computing Device Showdown: Users evaluated spatial computing devices like Meta Quest and Xreal for desktop use, debating their effectiveness in multi-monitor setups.
- While Meta Quest was favored for native application support, some limitations in optical quality were highlighted.
- GPT Integration Bug Report: Users noted that custom GPTs can no longer be integrated into another GPT's conversation with the '@' symbol, a shift that might indicate a bug.
- They suggested reaching out to support since some users still managed to use this function.
- Text2SQL Query Concerns: There's active discussion surrounding experiences with text2sql implementations, particularly in managing complex queries with LLMs.
- Users emphasized the need to keep context clear to avoid overwhelming outputs while fetching relevant data.
OpenRouter (Alex Atallah) Discord
- Inflection Models Go Live: The @inflectionAI model, powering @Pi, is now available on OpenRouter with no minimum spend and a playful focus on emojis 🍓.
- This model aims to enhance user interactions by allowing for a more engaging and fun chat experience with emoji integration 🤗.
- Grok 2 Launches with Exciting Access: OpenRouter now offers Grok 2 and Grok 2 Mini at a rate of $4.2/m input and $6.9/m output, though initially rate limited, as detailed in their announcement.
- Users appreciate the robust capabilities but note the critical nature of resource management during interactions.
- MythoMax Endpoint Offers Free Access: OpenRouter has launched a free MythoMax endpoint, broadening accessibility for users looking to leverage advanced models.
- This initiative aims to enhance user experience by providing more choices without additional costs.
- Chatroom Improvements Enhance Usability: Users can now drag and drop or paste images directly in chatrooms, enhancing overall interaction quality.
- These improvements reflect OpenRouter's commitment to streamlined and user-friendly communication within its platform.
- Grok API Running into Issues: Frequent errors like '500 Internal Server Error' and 'Rate limit exceeded' are reported by users encountering issues with the Grok API, which remains classified as experimental.
- It’s advised to consider beta models and other alternatives to mitigate these problems.
aider (Paul Gauthier) Discord
- Aider AI LLC Secures Source Code Home: The establishment of Aider AI LLC ensures that the aider source code is held under the Apache 2.0 license, maintaining its status as a completely free and open source project.
- “This is a community-driven effort with no funding rounds or employees involved,” reaffirming commitment to open-source principles.
- Users Overcome Aider Installation Challenges: Feedback indicated that installing Aider through
pipxdrastically simplified the setup process, avoiding lengthy install problems.- A user highlighted that adhering closely to the installation guide could mitigate installation issues.
- Jetbrains Plugin Crashes Spark Debate: Users reported that the Jetbrains plugin for Aider crashes on launch, pushing some to use Aider directly via terminal.
- Discussion focused on the plugin's lack of essential features, including file capture and keybindings, leading to frustration.
- Caution on Aider with Corporate Codebases: Concerns arose about using Aider on corporate codebases due to potential policy violations and risks of data leakage.
- While some emphasized that Aider operates locally without data sharing, worries about API use and screen sharing persisted.
- Comparative Performance of LLM Models: Debate over the effectiveness of various LLMs when integrated with Aider led to discussions about model performance, particularly regarding models like Grok-2 and GPT-4o.
- Members noted a need for careful selection of models to ensure optimal outputs in coding tasks.
Nous Research AI Discord
- Nous Research evolves into a startup: Nous Research, originally a Discord group, has transitioned into a funded startup focused on AI development, particularly in open-source projects.
- The community now plays a crucial role in facilitating collaboration and sharing ideas in the realm of AI research.
- DisTrO accelerates model training: DisTrO is designed to enable faster AI model training across the internet, promoting community-driven development as an alternative to closed models.
- The initiative aims to ensure sustained progress in the open-source field.
- Model collapse in neural networks revealed: A recent study examined the model collapse phenomenon, indicating that even 1% synthetic data can lead to significant performance decay.
- The research warns that larger models may exacerbate this collapse, challenging conventional scaling approaches.
- GSM-Symbolic improves LLM evaluations: The introduction of the GSM-Symbolic benchmark offers enhanced metrics for evaluating mathematical reasoning capabilities in LLMs.
- This benchmark diversifies assessment methods, promoting more reliable evaluations of language models.
- OpenAI faces scrutiny over Swarm framework: Accusations emerged alleging that OpenAI infringed on Kye Gomez's Swarms framework, with claims of stolen code and methodology.
- Potential legal actions are being considered unless investments are directed towards their project.
Perplexity AI Discord
- Reasoning Mode Launches in Pro Search: The Perplexity Team rolled out Reasoning Mode, an experimental feature that detects when extra compute can improve answers, encouraging users to share use cases in the feedback channel.
- Users provided various inquiries for Pro Search examples, including finding OpenAI co-founders and top-rated films, aiming to utilize the enhanced functionality.
- Navigating Cost Implications of AI Image Generation: Discussions surfaced regarding the costs tied to AI image generation, urging users to consider budgeting for these capabilities, more details found here.
- This dialogue highlighted the balance between affordability and project demands for high-quality visual output.
- User Frustrations with API Source URLs: Users grappled with the API not displaying source URLs in responses, prompting a call for support that met with silence, leaving inquiries unanswered.
- Discussion turned toward the online API, with references to
sonar-onlinemodels available at Perplexity API Docs, aiming to clarify model functionalities.
- Discussion turned toward the online API, with references to
- Mixed Feedback on AI Model Performance: Users expressed mixed experiences using various AI models, with some favoring Claude for coding tasks over recent updates impacting O1 mini performance.
- Concerns were raised about the Perplexity API's ability to deliver quality responses similar to online interactions, highlighting significant discrepancies.
- Exciting Updates in Perplexity Pro Features: Updates in Perplexity Pro sparked recent interest as users shared insights on new features aimed at enhancing engagement and functionality.
- Members can explore these changes further via this link, fueling active discussions on best practices.
GPU MODE Discord
- Attention Layer Implementation Confusion: A member seeks a tutorial for an Attention layer using cuDNN’s SDPA in Python, feeling lost on instantiating pygraph. They follow a notebook from the cudnn-frontend repository.
- Any help would be appreciated in clarifying the implementation details.
- Performance Profiling Discrepancies in PyTorch and Triton: Members find significant differences in performance results while profiling with PyTorch, Triton, and CUDA, leading to questions on which profiler to trust.
- Despite Triton claiming equal performance overall, self-evaluations seem to place PyTorch ahead in many tests.
- Metal Programming Challenges on Apple Silicon: Members report difficulties in using Docker with Apple Silicon GPUs, citing unresolved issues within the community. An internal ticket for the problem remains open without active work.
- Discussion also touches on a PR for the torch.special.i0 operator, focusing on MPS support enhancements.
- Entropix Sampling Skepticism: Skepticism flares around Entropix sampling, with some alleging it feels like nonsensical cult stuff, raising questions about its credibility.
- Despite concerns, a recent blog post mentions its aim to simplify reasoning without extensive modifications.
- Strategies for Efficient LLM Deployment: An online meetup is scheduled on October 5th at 4 PM PST to discuss LLM deployment, featuring contributions from SGLang, FlashInfer, and MLC LLM.
- Topics include low CPU overhead scheduling and kernel generation for performant LLM serving, with opportunities for community interaction.
Cohere Discord
- CohereForAI Contributions Encouraged: Members emphasized the importance of meaningful contributions to the CohereForAI community, suggesting citizen science as an entry point for those engaged in AI.
- One individual expressed a desire to contribute and mentor, aligning projects with a vision for a symbiotic relationship with technology.
- AI Innovators Win Nobel Prize: Sir John J. Hopfield and Sir Geoffrey E. Hinton received the 2024 Nobel Prize in Physics for their groundbreaking contributions to AI and neural networks.
- Their work has laid the foundational discoveries crucial for advancing machine learning technologies.
- API Tokens Confusion: A member questioned the necessity of using
<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>tokens for API requests, raising concerns about the quality of responses without them.- Will the responses still be decent without including these tokens? remains an unanswered question in the community.
- Clarification Needed on Cohere Rerank Pricing: There was confusion regarding whether web-search pricing is included in Cohere's Rerank pricing structure, as members couldn't find relevant details on the site.
- Understanding this pricing model is vital for planning effective implementation strategies.
- Upcoming Gen AI Hackathon Announcement: Members are invited to participate in the Gen AI Hackathon organized by CreatorsCorner, aiming to create innovative multi-agent systems.
- This event encourages collaboration to enhance human potential through intelligent solutions, as noted in the hackathon invite.
Modular (Mojo 🔥) Discord
- Mojo Installation Frustrations: Users are encountering installation issues with Mojo, leading to broken playground functionality and demo code errors, with a lack of clear tutorials available.
- One user noted that this disjointed process for Magic, Mojo, and MAX is causing significant confusion in the community.
- AES Hardware Support Advances: A member showcased their progress on implementing AES hardware support in Mojo via LLVM intrinsics, enhancing library integration capabilities.
- This effort reinforces Mojo's flexibility, allowing for advanced hardware functionalities to be incorporated smoothly into projects.
- Compilation Times Under Review for MAX: Compilation times for MAX are around 300 ms for graphs post-initial run and 500 ms for the first compilation, even for simple tasks.
- Discussion highlighted the importance of improving cache hit times to optimize performance during development.
- Implicit Conversions Create Debate: Implicit conversions in Mojo Lists raised questions among members, as adding an Int to a List seems possibly unintended due to existing constructors.
- An ongoing issue is tracking this behavior, which might complicate type handling in future implementations.
- Building with Mojo Faces Linking Errors: Users faced linking errors while attempting to build Mojo files, indicating potential missing libraries during the compilation process.
- Assistance included checking magic environment activation and proper installation protocol via command line.
Latent Space Discord
- OpenAI Swarm Launches Experimental Framework: OpenAI introduced Swarm, a lightweight library for building multi-agent systems, highlighting a stateless abstraction for managing agent interactions without relying on the Assistants API.
- It’s an experimental framework, aiming to facilitate easy understanding of agent roles and handoffs.
- Entropix Gains Popularity Among Members: Members engaged in an enthusiastic discussion about Entropix, providing an overview that shed light on its functionality and potential impacts.
- As interest grows, users are eager to see forthcoming evaluation features linked to the tool’s progression.
- RAG Techniques for Enhanced AI Performance: Discussion on RAG Techniques centered around a GitHub repository showcasing advanced methods for integrating retrieval with generative models.
- Participants aim to optimize performance, comparing frameworks like Haystack to custom solutions for specific use cases.
- Insights from Jensen on NVIDIA's Infrastructure: In a recent interview, Jensen discussed NVIDIA's full stack approach to AI infrastructure, underlining the imperative for accelerated computing in modern applications.
- His remarks reaffirmed generative AI’s transformative potential and indicated the continuous need for innovation in this space.
- Production AI Engineering Episode Highlights: The latest podcast episode covered insights into Production AI Engineering, focusing on the critical role of Evals in the industry.
- Experts declared Evals central to the landscape of LLM Ops, emphasizing the need for robust evaluation metrics as a growing priority.
Stability.ai (Stable Diffusion) Discord
- 3060ti Proves Competent in Stable Diffusion: Discussion highlighted the effectiveness of the 3060ti for Stable Diffusion, performing surprisingly well despite its 8GB VRAM limitations. Users cited Flux image generation as a testament to the GPU's capabilities.
- One user asserted that, with the right techniques, the 3060ti can handle demanding tasks in AI image generation efficiently.
- Lora Training Outshines Embedding: Participants debated the benefits of Lora training over embedding, asserting that Lora typically results in higher quality images. While embedding impacts only the text encoder, Lora allows for more nuanced diffusion model training.
- This detail sparked interest in deeper discussions about workflow adjustments for optimal image quality.
- Image Upscaling Techniques Under Scrutiny: The community compared Tiled Diffusion with Ultimate SD Upscale, noting each method serves distinct purposes—VRAM management vs. resolution enhancement. The merits of both techniques were extensively evaluated in ongoing projects.
- Users agreed that understanding when to apply each technique can significantly affect the outcome of image processing tasks.
- Image to 3D Model Generation Still Needs Work: The complexities of image to 3D model generation generated considerable discussion, as participants recognized the existing gaps in effective solutions. Multi-view inference techniques emerged as the most reliable methods currently available.
- Members expressed a collective need for innovation in this area, as the challenges remain significant.
- Looking for Help with Product Photo Integration: A member sought advice on integrating product photos into various backgrounds, emphasizing the need for high-quality results over basic compositing. Suggestions pointed towards leveraging Lora training to achieve better blending in final images.
- The conversation underscored the importance of advanced techniques in fulfilling specific visual demands.
LlamaIndex Discord
- LlamaIndex Hackathon Approaches: The upcoming LlamaIndex Hackathon this weekend invites participants to engage and innovate, with all relevant slides and resources shared for pre-hack preparation.
- Participants are encouraged to utilize these resources to ensure their projects are well-grounded.
- Building RAG Pipelines Simplified: Check out this video that walks through a basic RAG pipeline setup using LlamaIndex, highlighting essential workflows and components.
- The tutorial features a simple implementation with the router query technique for improved accuracy.
- Integrating Chat History with RouterQueryEngine: An inquiry about the RouterQueryEngine indicated an interest in incorporating chat history through the inclusion of all chat messages to enhance interaction dynamics.
- Workflow suggestions and examples were shared to facilitate better integration practices.
- Challenges with PDF Image Extraction: Users faced difficulties with extracting images from PDFs, frequently encountering unexpected ASCII characters in outputs, creating confusion.
- Guidance was sought to clarify data exports from parsed results, indicating a need for better documentation or support.
- Interest in Colpali within LlamaIndex: Questions arose about the potential for Colpali implementation in LlamaIndex, amid a noted gap in documentation.
- While full embedding isn't currently supported, community interest suggests that adding it as a reranker could be on the horizon.
tinygrad (George Hotz) Discord
- Tinygrad Type Annotations Discussion: The team evaluated three PRs for adding type annotations in Tinygrad, dismissing one due to performance concerns and another recognized contributor prioritized higher.
- A PR was rejected after failing tests, prompting worries about the practicality of merging such changes.
- Bounties Require Proven Contributors: George emphasized that contributors with several merged PRs are prioritized for bounty tasks, noting a new $200 bounty for parallel SHA3 implementation.
- This highlights the necessity for experience as a pre-requisite for tackling larger contributions.
- Challenges in SHA256 Implementation: A proposal for a complete SHA256 implementation in Tinygrad sparked talks about integrating parallel processing despite current design limitations.
- George showed interest in exploring parallel capabilities to optimize the implementation.
- DDPM Schedulers Thrive on Metal: A member introduced their own DDPM scheduler for training diffusion models on Metal, filling a gap in Tinygrad's resources.
- They are willing to collaborate with others needing support with this new tool.
- Addressing Tensor Gradient Axis Issues: The community debated solutions for resolving gradient axis mismatches in tensors, offering multiple approaches like axis alignment and resharding.
- Concerns were raised about the wastefulness of resharding as a solution.
Interconnects (Nathan Lambert) Discord
- OpenAI O1 Model Replication Advances: The initial report on replicating OpenAI's O1 model showcases a new 'journey learning' paradigm that enhances mathematical reasoning with 327 training samples, resulting in an 8% improvement in performance.
- Microsoft's Top AI Researcher Joins OpenAI: Reports confirm that Sebastien Bubeck, a prominent AI researcher from Microsoft, is moving to OpenAI, raising questions about the motivations behind such transitions amid lucrative AI roles. The Information highlights this significant career shift.
- The movement has created a stir, with industry colleagues humorously speculating on the implications for existing AI teams.
- Ex-OpenAI Employees Launching Startups: A staggering 1,700 startups are anticipated to be founded by former OpenAI employees, marking a significant surge in the AI startup ecosystem.
- This trend reflects a shift toward innovation and diversification within the field, producing potential new leaders in AI technology.
- Dario Amodei's Influential Work Gains Recognition: Machines of Loving Grace has been lauded for its compelling title and engaging content, stirring interest in AI's potential benefits for society.
- This growing discourse signals a shift towards positive perceptions of AI's future, moving away from fear-based narratives.
- Folding@Home's Early Influence in AI: Discussion arose around Folding@Home and its perceived underwhelming impact, with some members asserting it was ahead of its time despite its pioneering contributions to biological computing.
- The conversation also acknowledged the relevance of established methods like docking in drug discovery that seemed overshadowed during the Nobel discussions.
DSPy Discord
- Next.JS Voice Interview Prep Platform Launch: A member announced the development of a full stack Next.JS voice interview prep/testing platform aimed at enhancing interview preparation through voice interaction.
- This platform is expected to significantly improve the user experience during interview training.
- GraphIC Transforms ICL Selection: The paper introduces GraphIC, a technique using graph-based representations and Bayesian Networks to enhance the selection of in-context examples for LLMs.
- It resolves limitations of text-based embedding methods in multi-step reasoning tasks by filtering out shallow semantics.
- LLM Classifier Seeks Ambiguity Handling: A user is working on training an LLM classifier and is seeking community input on handling classification ambiguities to effectively manage uncertain outputs.
- The suggestion involves adding a second output field in the LLM signature to declare ambiguities instead of creating separate classes.
- Assessing Output Effectiveness with Cosine Similarity: A member inquired about a metric to evaluate if chatbot outputs meet established criteria, considering cosine similarity to compare input queries with generated categories.
- Stuart is actively seeking suggestions to refine this approach for better off-topic detection.
- FastAPI Route Creation for Signatures: A member shared a code snippet that enables any dspy.Signature to turn into a FastAPI route, returning the predictor as a dictionary with init_instant function used for environment initialization.
- This implementation streamlines the request processing essential for developing APIs with DSPy.
LAION Discord
- LLaMA 3.2 Dominates Pop Culture Knowledge: LLaMA 3.2 excels over competitors with its training on 5 billion images, allowing for more coherent captions.
- In comparisons, LLaMA 3.2 demonstrates significant contextual understanding compared to models like Molmo and PixTral.
- PixTral Shines in Adult Content Scenarios: Members highlighted that PixTral stands out when focused on adult content, unlike LLaMA 3.2, which is better suited for broader contexts.
- The contrast indicates that while PixTral has a niche, LLaMA 3.2 maintains cultural relevance across more general applications.
- Epic Games Removing Sketchfab Sparks Concerns: Epic Games' removal of Sketchfab is set to eliminate 800k 3D models from Objaverse, prompting users to download urgently.
- This decision has raised alarms about its effects on the 3D modeling community and users relying on those resources.
- o1-mini Can't Compete with o1-preview: Reports indicate that o1-mini is outperformed by o1-preview, being described as brittle on straightforward tasks according to recent insights.
- Despite earlier claims of matching larger models, evidence suggests o1-preview excels even on Olympiad level tasks.
- Challenges with CLIP's Contrastive Training: Using CLIP for training T2I models speeds up the process but introduces artifacts tied to its contrastive training methods.
- These artifacts raise concerns about impacting overall training quality, suggesting trade-offs in efficiency and performance.
LLM Agents (Berkeley MOOC) Discord
- Graham Neubig's Lecture on AI Agents: Graham Neubig's lecture today at 3:00 PM PST discusses
- Neubig also highlights complexities in developing AI agents for large repositories, addressing issues such as file selection and integration of web browsing into workflows.
- Registration Delays and Troubles: Members confirmed that course registration remains open until Dec 12, with a successful sign-up experience shared after troubleshooting a linking issue.
- Participants reported challenges with Google Forms for quizzes, suggesting clearing browser cache resolved access issues; course timings are set for 3:00 PM to 5:00 PM PST.
- Defining the AI Agent: An AI agent autonomously performs tasks via interactions with APIs and databases, with ChatGPT classified as an agent, unlike gpt-3.5-turbo, which lacks such capabilities.
- Discussion also includes ongoing efforts to refine AI agent definitions, emphasizing the importance of community input via platforms like Twitter.
- Chain of Thought Enhances LLM Problem Solving: The Chain of Thought (CoT) methodology assists LLMs in breaking down complex tasks into manageable steps, promoting clarity in problem-solving.
- Members recognized CoT’s effectiveness through an example involving Apple, showcasing how systematic breakdowns lead to final solutions.
- AI-Powered Search Book Gains Attention: A member recommended this book as the go-to resource for AI-powered search, praising its anticipated impact over the coming years.
- The book is expected to serve as a vital reference for both AI practitioners and researchers, highlighting its future relevance in the field.
Torchtune Discord
- Gemma 2 Support leverages Flex Attention: Discussion focused on implementing Gemma 2 using Flex Attention, with logit softcapping identified as the main blocker that needs a proper
score_modfunction.- Members believe the tradeoff for Flex simplifies the process, although it may require high compute capabilities with CUDA.
- Introduction of the Aria Model: Aria, a new open multimodal AI model, showcases a 3.9B and 3.5B parameter architecture and excels in language and coding tasks, outperforming Pixtral-12B.
- While there are no direct benchmark comparisons yet, early indications show Aria’s capabilities surpass its contemporaries.
- LegoScale Revolutionizes Distributed Training: LegoScale introduces a customizable, PyTorch-native system for 3D parallel pre-training of large language models, significantly boosting performance.
- Its modular approach aims to simplify complex training across GPUs, potentially changing the landscape of distributed training.
- Insights from the State of AI 2024 Report: The State of AI Report 2024 by Nathan Benaich outlines significant trends and investment areas in AI with few mentions of models like Torchtune.
- This report serves to prompt discussions on the future of AI, particularly regarding its applications in medicine and biology.
- Out-of-Shared-Memory Issues with Flex Attention: A GitHub issue shared problems with flex attention on the RTX 4090, detailing errors linked to out-of-shared-memory problems.
- The conversation included a minimal reproduction code snippet, fostering collaboration for troubleshooting.
LangChain AI Discord
- Swarm.js Launch for Node.js Enthusiasts: Swarm.js, a lightweight Node.js SDK, orchestrates multi-agent systems using the OpenAI API, enabling seamless agent management and task execution.
- Developers can easily start by running
npm install openai-swarm-node, with the project actively inviting contributions and collaboration from both beginners and experts.
- Developers can easily start by running
- Community Closure Announcement: Jess announced that the LangChain Discord community is set to close on October 31, 2024, to focus on creating a new community platform.
- All members are encouraged to fill out a form for updates and provide feedback at community@langchain.dev, with invitations extended for potential moderators.
- Exploring Contextual Retrieval Techniques: A new YouTube video illustrates how to implement contextual retrieval using LangChain and OpenAI’s Swarm Agent, guiding viewers through the integration process.
- This informative content is aimed at enhancing information retrieval, making it especially relevant for those employing LangChain in their projects.
- bootstrap-rag v0.0.9 Goes Live!: bootstrap-rag v0.0.9 has been released with critical bug fixes, improved documentation, and integration with LangChain and MLflow-evals.
- The update also includes templates for Qdrant, enhancing retrieval-augmented generation capabilities, a key area for AI engineers focusing on efficient data handling.
- LangGraph Tutorial for Job Seekers: A new tutorial demonstrates building a two-node LangGraph app that analyzes resumes against job descriptions, offering practical benefits for job applicants. Watch here.
- The app can craft tailored cover letters and generate job-specific interview questions, making it a handy tool for those new to LangGraph.
OpenAccess AI Collective (axolotl) Discord
- Exploring Instruction Data Impact: A member raised a question about the use of instruction data during pretraining, emphasizing the potential benefits for model engagement.
- This topic could lead to a discussion on innovative pretraining techniques that enhance model adaptability.
- Config Sharing Sparks Suggestions: In a discussion on config sharing, a member requested a specific config while suggesting sample packing as a crucial update.
- Challenges with multi-GPU setups were highlighted, emphasizing the need for a thorough setup review.
- Fine-Tuning with Adapters: There was discussion on merging an existing Llama 3 adapter with a fine-tuned model to improve accuracy in tasks.
- A GitHub guide was shared for the merging process, reiterating the importance of proper config settings.
- Text Completion Training Enhances Instruction Models: Training instruct models like GPT-3.5-Instruct on text completion tasks can lead to marked performance improvements in instruction compliance.
- Community members cautioned about overfitting risks, advising diverse datasets for optimal training outcomes.
- Diversity Key to Avoiding Overfitting: Concerns about overfitting emerged when discussing training datasets, with a call for more diversity to enhance generalization.
- Members stressed monitoring performance across tasks to mitigate risks of degradation on unfamiliar datasets.
OpenInterpreter Discord
- Message Format Compliance Reminder: Members were reminded to adhere to the prescribed message format in the channel to maintain organization and clarity.
- The reminder emphasized the importance of following established guidelines to enhance channel communication.
- Reference to Channel Rules: A reference was made to existing rules guiding behavior and contributions within discussions in the channel.
- Members were encouraged to review these rules for better channel dynamics.
- Aria Becomes the Top Multimodal Model: Aria, by @rhymes_ai_, is now ranked first on the 🤗 Open LLM Leaderboard, showcasing 24.9B parameters and handling image, video, and text inputs.
- Users Praise Aria's Multimodal Capabilities: Users are enthusiastic about the 25.3B multimodal Aria model, calling it the BEST vision language model I have ever tried!
- Released under the Apache-2.0 license, fine-tuning scripts are also available for community engagement.
- AI Reasoning Capabilities Debated: A YouTube video titled 'Apple DROPS AI BOMBSHELL: LLMS CANNOT Reason' raises critical questions about language models' reasoning capabilities.
- The creator stimulates a dialogue around current AI limitations, urging viewers to prepare for AGI advancements.
AI21 Labs (Jamba) Discord
- Member Starts Support Inquiry on Jamba: A user initiated a support thread about Jamba issues, asking if the channel was appropriate for assistance.
- Another member confirmed they had addressed the inquiry in the same channel, promoting ongoing discussions there.
- Importance of Thread Continuity for Jamba Issues: In response to the support inquiry, a member emphasized keeping the discussion within the original thread to maintain clarity.
- They pointed out that this approach would facilitate easier access to pertinent information in the future.
Mozilla AI Discord
- Panel Discussion on Community Engagement Strategies: A panel of Community and Developer Relations experts, including Jillian Bejtlich and Rynn Mancuso, is set to discuss actionable strategies for enhancing community engagement during the upcoming event.
- This engagement-focused session aims to equip attendees with practical techniques for growing user bases and boosting contributions to projects.
- Tactical Insights for Building Thriving Communities: The panel will share tactical advice on how to cultivate a successful community around projects, emphasizing the importance of relationship-building beyond coding.
- Project leads seeking to refine their community-building skills will find this session particularly useful for networking and strategy enhancement.
- RSVP for the Community Panel Discussion!: Participants are encouraged to RSVP for the panel discussion on community-building practices here to secure their spot.
- “Don’t miss out on this invaluable opportunity!” resonates within the channel, urging engagement from the community.
DiscoResearch Discord
- Introducing the Backtrack Sampler: Mihai4256 shared an interesting GitHub repository focused on a backtracking sampling method.
- This repository is likely to attract AI engineers interested in advanced sampling techniques and model optimization.
- Check Out the GitHub Repo: The repository provides innovative approaches to sampling that could enhance algorithm efficiency and model accuracy.
- Mihai4256 encourages collaboration and feedback from the community on the implementation discussed in the repo.
Gorilla LLM (Berkeley Function Calling) Discord
- Performance Struggles in Multi-Turn Evaluations: Members encountered a ~0% performance rate during multi-turn evaluations, as the model fails to exit the count loop despite correct predictions.
- The discussion highlighted various efforts to find a viable solutions for this evaluation conundrum.
- Workaround Boosts Multi-Turn Evaluation Ratings: A temporary code modification in base_handler.py improved evaluation accuracy to ~15% performance by attempting each round only once.
- However, the need for compliance with modification restrictions has left members seeking alternative strategies to enhance performance.
The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The LLM Finetuning (Hamel + Dan) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
PART 2: Detailed by-Channel summaries and links
The full channel by channel breakdowns have been truncated for email.
If you want the full breakdown, please visit the web version of this email: !
If you enjoyed AInews, please share with a friend! Thanks in advance!