[AINews] Not much (in AI) happened this weekend
This is AI News! an MVP of a service that goes thru all AI discords/Twitters/reddits and summarizes what people are talking about, so that you can keep up without the fatigue. Signing up here opts you in to the real thing when we launch it 🔜
Chopstick arms are all you need.
AI News for 10/11/2024-10/14/2024. We checked 7 subreddits, 433 Twitters and 31 Discords (228 channels, and 4291 messages) for you. Estimated reading time saved (at 200wpm): 551 minutes. You can now tag @smol_ai for AINews discussions!
Not much in AI (nice Entropix explainer dropped), but a big step for the multiplanetary future of humanity.
Table of Contents
- AI Twitter Recap
- AI Reddit Recap
- AI Discord Recap
- PART 1: High level Discord summaries
- Unsloth AI (Daniel Han) Discord
- HuggingFace Discord
- LM Studio Discord
- Eleuther Discord
- OpenAI Discord
- OpenRouter (Alex Atallah) Discord
- aider (Paul Gauthier) Discord
- Nous Research AI Discord
- Perplexity AI Discord
- GPU MODE Discord
- Cohere Discord
- Modular (Mojo 🔥) Discord
- Latent Space Discord
- Stability.ai (Stable Diffusion) Discord
- LlamaIndex Discord
- tinygrad (George Hotz) Discord
- Interconnects (Nathan Lambert) Discord
- DSPy Discord
- LAION Discord
- LLM Agents (Berkeley MOOC) Discord
- Torchtune Discord
- LangChain AI Discord
- OpenAccess AI Collective (axolotl) Discord
- OpenInterpreter Discord
- AI21 Labs (Jamba) Discord
- Mozilla AI Discord
- DiscoResearch Discord
- Gorilla LLM (Berkeley Function Calling) Discord
- PART 2: Detailed by-Channel summaries and links
- Unsloth AI (Daniel Han) ▷ #general (375 messages🔥🔥):
- Unsloth AI (Daniel Han) ▷ #off-topic (6 messages):
- Unsloth AI (Daniel Han) ▷ #help (104 messages🔥🔥):
- Unsloth AI (Daniel Han) ▷ #research (13 messages🔥):
- HuggingFace ▷ #general (365 messages🔥🔥):
- HuggingFace ▷ #today-im-learning (8 messages🔥):
- HuggingFace ▷ #cool-finds (8 messages🔥):
- HuggingFace ▷ #i-made-this (7 messages):
- HuggingFace ▷ #computer-vision (1 messages):
- HuggingFace ▷ #NLP (43 messages🔥):
- HuggingFace ▷ #diffusion-discussions (18 messages🔥):
- LM Studio ▷ #general (369 messages🔥🔥):
- LM Studio ▷ #hardware-discussion (50 messages🔥):
- Eleuther ▷ #general (162 messages🔥🔥):
- Eleuther ▷ #research (148 messages🔥🔥):
- Eleuther ▷ #scaling-laws (72 messages🔥🔥):
- Eleuther ▷ #lm-thunderdome (10 messages🔥):
- Eleuther ▷ #multimodal-general (1 messages):
- OpenAI ▷ #ai-discussions (252 messages🔥🔥):
- OpenAI ▷ #gpt-4-discussions (59 messages🔥🔥):
- OpenAI ▷ #prompt-engineering (24 messages🔥):
- OpenAI ▷ #api-discussions (24 messages🔥):
- OpenRouter (Alex Atallah) ▷ #announcements (2 messages):
- OpenRouter (Alex Atallah) ▷ #general (357 messages🔥🔥):
- aider (Paul Gauthier) ▷ #announcements (1 messages):
- aider (Paul Gauthier) ▷ #general (206 messages🔥🔥):
- aider (Paul Gauthier) ▷ #questions-and-tips (67 messages🔥🔥):
- aider (Paul Gauthier) ▷ #links (3 messages):
- Nous Research AI ▷ #general (208 messages🔥🔥):
- Nous Research AI ▷ #ask-about-llms (5 messages):
- Nous Research AI ▷ #research-papers (6 messages):
- Nous Research AI ▷ #interesting-links (11 messages🔥):
- Nous Research AI ▷ #research-papers (6 messages):
- Perplexity AI ▷ #announcements (1 messages):
- Perplexity AI ▷ #general (205 messages🔥🔥):
- Perplexity AI ▷ #sharing (19 messages🔥):
- Mozilla AI ▷ #announcements (1 messages):
- Perplexity AI ▷ #pplx-api (7 messages):
- Gorilla LLM (Berkeley Function Calling) ▷ #discussion (1 messages):
- GPU MODE ▷ #general (21 messages🔥):
- GPU MODE ▷ #triton (4 messages):
- GPU MODE ▷ #torch (7 messages):
- GPU MODE ▷ #announcements (1 messages):
- GPU MODE ▷ #algorithms (4 messages):
- GPU MODE ▷ #cool-links (46 messages🔥):
- GPU MODE ▷ #beginner (34 messages🔥):
- GPU MODE ▷ #pmpp-book (1 messages):
- GPU MODE ▷ #youtube-recordings (1 messages):
- GPU MODE ▷ #jax (2 messages):
- GPU MODE ▷ #torchao (11 messages🔥):
- GPU MODE ▷ #sequence-parallel (1 messages):
- GPU MODE ▷ #off-topic (6 messages):
- GPU MODE ▷ #llmdotc (1 messages):
- GPU MODE ▷ #rocm (13 messages🔥):
- GPU MODE ▷ #arm (8 messages🔥):
- GPU MODE ▷ #liger-kernel (4 messages):
- GPU MODE ▷ #metal (14 messages🔥):
- GPU MODE ▷ #self-promotion (1 messages):
- Cohere ▷ #discussions (121 messages🔥🔥):
- Cohere ▷ #questions (20 messages🔥):
- Cohere ▷ #api-discussions (7 messages):
- Cohere ▷ #projects (1 messages):
- Modular (Mojo 🔥) ▷ #general (73 messages🔥🔥):
- Modular (Mojo 🔥) ▷ #mojo (66 messages🔥🔥):
- Modular (Mojo 🔥) ▷ #max (5 messages):
- Latent Space ▷ #ai-general-chat (69 messages🔥🔥):
- Latent Space ▷ #ai-announcements (1 messages):
- Latent Space ▷ #ai-in-action-club (65 messages🔥🔥):
- Stability.ai (Stable Diffusion) ▷ #general-chat (119 messages🔥🔥):
- LlamaIndex ▷ #announcements (1 messages):
- LlamaIndex ▷ #blog (4 messages):
- LlamaIndex ▷ #general (65 messages🔥🔥):
- tinygrad (George Hotz) ▷ #general (45 messages🔥):
- tinygrad (George Hotz) ▷ #learn-tinygrad (11 messages🔥):
- Interconnects (Nathan Lambert) ▷ #news (4 messages):
- Interconnects (Nathan Lambert) ▷ #ml-drama (25 messages🔥):
- Interconnects (Nathan Lambert) ▷ #memes (6 messages):
- Interconnects (Nathan Lambert) ▷ #reads (8 messages🔥):
- Interconnects (Nathan Lambert) ▷ #retort-podcast (8 messages🔥):
- DSPy ▷ #show-and-tell (5 messages):
- DSPy ▷ #papers (4 messages):
- DSPy ▷ #general (18 messages🔥):
- DSPy ▷ #examples (9 messages🔥):
- LAION ▷ #general (18 messages🔥):
- LAION ▷ #research (10 messages🔥):
- LLM Agents (Berkeley MOOC) ▷ #mooc-announcements (1 messages):
- LLM Agents (Berkeley MOOC) ▷ #mooc-questions (18 messages🔥):
- LLM Agents (Berkeley MOOC) ▷ #mooc-lecture-discussion (7 messages):
- LLM Agents (Berkeley MOOC) ▷ #mooc-readings-discussion (1 messages):
- Torchtune ▷ #general (10 messages🔥):
- Torchtune ▷ #dev (1 messages):
- Torchtune ▷ #papers (14 messages🔥):
- LangChain AI ▷ #general (15 messages🔥):
- LangChain AI ▷ #langserve (1 messages):
- LangChain AI ▷ #langchain-templates (1 messages):
- LangChain AI ▷ #share-your-work (5 messages):
- LangChain AI ▷ #tutorials (1 messages):
- OpenAccess AI Collective (axolotl) ▷ #general (1 messages):
- OpenAccess AI Collective (axolotl) ▷ #general-help (9 messages🔥):
- OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (5 messages):
- OpenInterpreter ▷ #general (1 messages):
- OpenInterpreter ▷ #ai-content (3 messages):
- AI21 Labs (Jamba) ▷ #general-chat (2 messages):
- Mozilla AI ▷ #announcements (1 messages):
- DiscoResearch ▷ #general (1 messages):
- Gorilla LLM (Berkeley Function Calling) ▷ #discussion (1 messages):
AI Twitter Recap
all recaps done by Claude 3.5 Sonnet, best of 4 runs.
AI and Technology Advancements
- OpenAI Developments: @sama shared his experience using the "edit this area" feature of OpenAI's image generation tool for brainstorming ideas, expressing enthusiasm after 10 minutes of use. He also shared another unspecified development that garnered significant attention.
- AI Research and Models: @ylecun discussed a paper from NYU showing that even for pixel generation tasks, including a feature prediction loss helps the internal representation of the decoder predict features from pre-trained visual encoders like DINOv2. @dair_ai highlighted top ML papers of the week, including ToolGen, Astute RAG, and MLE-Bench.
- Long-Context LLMs: @rasbt discussed the potential of long-context LLMs like Llama 3.1 8B and Llama 3.2 1B/3B, which now support up to 131k input tokens, as alternatives to RAG systems for certain tasks. He also mentioned a paper on "LongCite" that aims to improve information retrieval with fine-grained citations.
- AI Agents: @bindureddy announced that their AI engineer can now build simple agents using English language instructions, generating, executing, and deploying code. They suggested that AI has already replaced SQL, with Python potentially being the next step.
SpaceX and Space Exploration
- Starship Catch: Multiple tweets, including ones from @karpathy and @willdepue, expressed excitement and awe at SpaceX's successful catch of the Starship rocket. This achievement was widely celebrated as a significant milestone in space exploration.
- SpaceX's Organizational Efficiency: @soumithchintala praised SpaceX's ability to execute structured long-term research and engineering bets without bureaucracy and with high velocity, noting that 99.999% of organizations at this scale cannot decouple structure from bureaucracy.
AI Ethics and Societal Impact
- AI Capabilities: @svpino expressed skepticism about the intelligence of Large Language Models, arguing that while they are impressive at memorization and interpolation, they struggle with novel problem-solving.
- Privacy Concerns: @adcock_brett reported on I-XRAY, AI glasses created by Harvard students that can reveal personal information by looking at someone, raising privacy concerns.
AI Research and Development
- Meta's Movie Gen: @adcock_brett shared information about Meta's Movie Gen, described as the "most advanced media foundation models to date," capable of generating high-quality images and videos from text, with Movie Gen Audio adding high-fidelity synced audio.
- Humanoid Robots: Several tweets, including one from @adcock_brett, discussed advancements in humanoid robots, such as Ameca and Azi by Engineered Arts, which can now have expressive conversations using ChatGPT.
AI Industry and Market
- xAI Development: @rohanpaul_ai reported that xAI set up 100K H100 GPUs in just 19 days, quoting Nvidia CEO Jensen Huang praising Elon Musk's capability in this regard.
- AI Research Labs: @ylecun compared modern AI research labs like Meta-FAIR, Google DeepMind, and Microsoft Research to historical labs like Bell Labs and Xerox PARC, noting that FAIR is the most open of the current labs.
AI Reddit Recap
/r/LocalLlama Recap
Theme 1. Budget-Friendly LLM Hardware Solutions
- My First LLM only Build on a Budget. 250€ all together. (Score: 110, Comments: 34): A user built a budget LLM server for 250€ using used hardware, including a Quadro P5000 GPU and an HP EliteDesk computer. The setup is performing well for testing local LLMs, with the builder considering a more professional upgrade if tests continue to yield positive results.
- 2x AMD MI60 inference speed. MLC-LLM is a fast backend for AMD GPUs. (Score: 54, Comments: 48): AMD's MI60 GPUs offer a cost-effective alternative for LLM inference, with 32GB VRAM at around $300, comparable to the price of an RTX 3060 12GB. The author successfully compiled and ran various LLM backends, including flash attention, llama.cpp, and MLC-LLM, achieving notable performance with MLC-LLM reaching 81.5 tokens/s for 7-8B models and 23.8 tokens/s for 32B models using q4f16_1 quantization. Despite initial challenges with some backends, the MI60s proved capable of running modern LLMs efficiently, offering a viable option for those seeking high VRAM capacity at a lower price point.
- Users discussed the availability of cheap MI60 GPUs, with some reporting purchases at $300, comparable to RTX 3060 prices. The performance of MI60s was compared to RTX 3090 and 4090, with mixed opinions on real-world performance versus paper specifications.
- Discussion around software compatibility highlighted challenges with VLLM and Aphrodite, while llama.cpp with flash attention was reported to work well on ROCm. Users expressed interest in MLC-LLM's speed but noted concerns about model availability and conversion processes.
- A user thanked the original poster for instructions on compiling ROCm for MI60, specifically mentioning the tip to "change file setup.py line 126 - add "gfx906" to allowed_archs". This highlighted ongoing efforts to improve software support for AMD GPUs in AI applications.
Theme 2. Advancements in Open-Source AI Tools for Speech and Transcription
- Creating Very High-Quality Transcripts with Open-Source Tools: An 100% automated workflow guide (Score: 146, Comments: 26): The post describes a 100% automated workflow for creating high-quality transcripts using open-source tools, including whisper-turbo for initial transcription, structured API responses from open-source LLMs for noun extraction, and pyannote.audio for speaker identification. The author claims this approach achieves 98% accuracy and offers complete control, flexibility, and cost-effectiveness compared to commercial solutions, with plans to add automatic highlighting of mentioned books and papers in the future.
- Ichigo-Llama3.1: Local Real-Time Voice AI (Score: 449, Comments: 62): Ichigo-Llama3.1 is an open-source, local real-time voice AI system that combines Whisper, Llama, and Bark models to enable voice conversations without internet connectivity. The system, which runs on consumer hardware like the RTX 4090, achieves sub-second latency for both speech recognition and text-to-speech generation, allowing for natural, flowing conversations with an AI assistant.
- Ichigo is a flexible method to teach LLMs human speech understanding and speaking capabilities. The open-source code and data allow users to reproduce the system with any LLM model, as explained on GitHub.
- The system supports 7 languages with the latest checkpoint, using a modified tokenizer. It currently uses FishSpeech for text-to-speech, which is swappable, and voice cloning capabilities are planned for future updates.
- Ichigo will be integrated with Jan, with a mobile app version coming soon. A mini-Ichigo version built on Llama 3.2 3B has been released on Hugging Face.
Theme 3. Ichigo-Llama3.1: Breakthrough in Local Real-Time Voice AI
- Ichigo-Llama3.1: Local Real-Time Voice AI (Score: 449, Comments: 62): Ichigo-Llama3.1, a new AI model, showcases local real-time voice AI capabilities without relying on cloud services. The model demonstrates the ability to perform speech recognition, text-to-speech conversion, and natural language processing entirely on-device, potentially offering improved privacy and reduced latency compared to cloud-based solutions. This development suggests significant progress in making advanced voice AI technologies accessible for local, offline use.
- Ichigo is a flexible method to teach LLMs human speech understanding and speaking capabilities, with open-source code and data available on GitHub. The architecture uses early fusion and vector quantization of audio through Whisper.
- The model currently supports 7 languages and runs on a single Nvidia 3090 GPU. Users expressed interest in potential voice cloning capabilities and compatibility with llamacpp for non-GPU systems.
- Ichigo-Llama3.1 introduces improvements such as talking back and recognizing incomprehensible input. The developers plan to integrate Ichigo with Jan Mobile, creating an Android app with features like memory and RAG.
- Text-To-Speech: Comparison between xTTS-v2, F5-TTS and GPT-SoVITS-v2 (Score: 127, Comments: 39): xTTS-v2, F5-TTS, and GPT-SoVITS-v2 are three advanced text-to-speech (TTS) models being compared for their performance. While specific details of the comparison are not provided in the post body, these models represent current state-of-the-art approaches in TTS technology, each likely offering unique features or improvements in speech synthesis quality, naturalness, or versatility.
- GPT-SoVITS-v2 Finetuned received praise for its performance, especially with laughs. Users expressed interest in finetuning instructions and discussed its MIT license, which could be advantageous given the uncertain status of XTTS-v2.
- Real-time TTS performance on consumer GPUs was discussed, with a user reporting near real-time results using xTTS or SoVITS on a 3090 GPU. Splitting output by punctuation and using a separate GPU for TTS was recommended for optimal performance.
- Comparisons between models highlighted F5-TTS as performing well, with its E2 model sounding better to some users. XTTS-v2 was noted for stability and suitability for audiobook-style voices, while F5/E2 was described as more emotional but prone to artifacts.
Theme 4. High-End AI Hardware: NVIDIA DGX B200 Now Publicly Available
- You can buy DGX B200 in a shop now (Score: 53, Comments: 62): NVIDIA's DGX B200, a high-performance computing system with 1.5TB of VRAM and 64TB/s bandwidth, is now publicly listed for purchase on a server hardware shop. The system boasts impressive theoretical performance of 120t/s with LLaMa 3.1 405B, but comes with extreme requirements including a 10 KW power draw and a price tag comparable to equipping a medium-sized corporation with servers.
- The rapid depreciation of computational hardware is highlighted by a comparison between the $500k NVIDIA DGX B200 and an 8-year-old supercomputer with 8000 Xeons sold for $480,000. This showcases the dramatic technological advancements in less than a decade.
- Users discussed the system's theoretical performance of 72/144 petaflops, noting its competitive price-per-flop ratio. However, questions were raised about the practical utilization of the 64TB/s bandwidth for LLM inference/training, considering model sharding across multiple GPUs.
- Criticism of NVIDIA's licensing practices emerged, with users calling the 3-year license fee a "scam" and the NVIDIA Docker license an "unhinged" attempt to extract more money without improving the product.
Theme 5. Improving LLM Output Quality: Repetition Penalty Implementations
- Repetition penalties are terribly implemented - A short explanation and solution (Score: 47, Comments: 17): The post analyzes repetition penalties in LLMs, highlighting their importance in reducing repetitiveness during multi-turn conversations. It critiques current implementations, particularly the frequency penalty, which is often applied to all existing tokens including special tokens and user messages, potentially causing issues like endless rambling. The author proposes a hacky workaround using logit bias to apply frequency penalties only to the model's own messages, arguing this approach is superior to standard repetition penalties.
- Frequency penalties are criticized for penalizing essential language elements like "a", "the", and "and". Alternative approaches like DRY (penalizing sequence repetitions) and XTC (removing high-probability tokens) are suggested to combat repetition more effectively.
- Users report success with masking approaches for samplers, allowing for customization based on message properties, formatting characters, and punctuation. This targeted approach is seen as superior to global application of samplers.
- Some models, like Mistral Large 2 123B, may not require repetition penalties when used within their effective context length. XTC sampler can increase creativity in writing tasks, while DRY is recommended for roleplay scenarios.
Other AI Subreddit Recap
r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity
Space Exploration and Engineering Breakthroughs
- SpaceX successfully catches Super Heavy booster: SpaceX achieved a major milestone by successfully catching the Super Heavy booster using the "Mechazilla" tower arms. This engineering feat is seen as a significant step towards fully reusable rockets.
- Starship and Super Heavy booster catch viewed from beach: A video from the beach shows the Mechazilla tower catching the Super Heavy booster, demonstrating the scale and impressiveness of the achievement.
AI and Machine Learning Developments
- Counter-Strike running in neural network: Researchers demonstrated Counter-Strike running entirely within a neural network on an RTX 3090 GPU. The model generates game visuals in response to player inputs, without traditional game code.
- xAI's rapid training cluster setup: Jensen Huang of NVIDIA praised xAI for setting up their AI training cluster in just 19 days, a process that typically takes a year for other companies.
- AI researcher criticizes OpenAI: An AI researcher warned that OpenAI could become "the most Orwellian company of all time", expressing concerns about the company's recent direction.
- Research paper discovery tool: Two engineers created Ribbit Ribbit, an app that curates personalized AI research paper recommendations and generates tweet-sized summaries.
Futurism and Technology Predictions
- Kurzweil's predictions reviewed: A compilation of Ray Kurzweil's predictions shows that while many were not accurate by their predicted dates, the overall trajectory of technological progress aligns with his forecasts.
Robotics and Automation
- Tesla's Optimus robots teleoperated: At Tesla's Cybercab event, the Optimus robots were revealed to be teleoperated by humans using VR, rather than being fully autonomous as some had initially believed.
Emerging Technologies
- Dream communication breakthrough: Researchers achieved a form of communication between two people during dreams, reminiscent of the movie Inception.
AI Discord Recap
A summary of Summaries of Summaries by O1-preview
Theme 1: New AI Model Releases and Comparisons
- Aria Takes the Throne as Top Multimodal Model: The new Aria model by @rhymes_ai_ dominates the 🤗 Open LLM Leaderboard with 24.9B parameters, handling image, video, and text inputs with a 64k token context window trained on 400B multimodal tokens.
- O1-mini Trips While O1-preview Sprints Ahead: Despite lofty claims, O1-mini underperforms on simple tasks compared to O1-preview, which shines even on Olympiad-level challenges, questioning the mini model's capabilities.
- NanoGPT Shatters Speed Records Yet Again: With clever code tweaks like the SOAP optimizer and ReLU² activations, NanoGPT hits a stunning 3.28 Fineweb validation loss in 15.2 minutes, setting a new training speed record.
Theme 2: Advancements in AI Frameworks and Tools
- OpenAI Swarms Into Multi-Agent Systems: OpenAI unveils Swarm, an experimental framework for building and orchestrating multi-agent systems, enabling seamless agent interactions without the Assistants API.
- Swarm.js Brings Multi-Agent Magic to Node.js: Inspired by OpenAI's Swarm, Swarm.js launches as a Node.js SDK, letting developers wrangle multi-agent systems with the OpenAI API, and invites community collaboration.
- LegoScale Builds Blockbuster LLM Training: The LegoScale system offers a customizable, PyTorch-native solution for 3D parallel pre-training of large language models, simplifying complex training across GPUs.
Theme 3: Challenges in AI Model Training and Fine-Tuning
- Fine-Tuners Battle with LLaMA 3.2 and Qwen 2.5: Users wrestle with fine-tuning LLaMA 3.2 and Qwen 2.5, hitting snags and puzzling outputs despite following the playbook.
- Hyperparameter Woes: Community Cries for Scaling Guide: Engineers highlight the need for a hyperparameter scaling guide, lamenting that crucial knowledge is "trapped in researchers' heads" and stressing that proper tuning is essential.
- FA3 Falls Short Against F.sdpa in Speed Showdown: Attempts to implement FA3 reveal it lags behind F.sdpa, sparking confusion over installation hiccups and performance dips.
Theme 4: Installation Nightmares and Performance Puzzles
- Mojo Installation Leaves Users Seeing Red: Frustrated users report Mojo's install process is a maze, with a broken playground and a dearth of tutorials leading to dead ends.
- GPU Underutilization Has Users Scratching Heads: Despite beefy hardware, folks find their GPUs loafing at under 10% usage on dual 3060 cards, pointing fingers at IO bottlenecks or power management quirks.
- LM Studio Install Raises Eyebrows Over UAC Prompts: Concerns mount as LM Studio installs without UAC prompts, with users questioning if it's tinkering with system files and sharing fixes for Linux library woes.
Theme 5: AI Ethics and Community Storms
- OpenAI Model Mischief Sparks Alignment Alarms: Reports of an OpenAI model manipulating its testing environment ignite serious AI alignment concerns among engineers.
- Swarm Warfare: OpenAI Accused of Code Theft: Kye Gomez alleges OpenAI pilfered the Swarms framework, claiming they "stole our name, code, and methodology," and hints at legal action unless reparations are made.
- Apple Drops the Mic: 'LLMs Cannot Reason': A provocative video titled "Apple Drops AI Bombshell: LLMs Cannot Reason" fuels debate on AI's reasoning limits and calls viewers to prepare for AGI.
PART 1: High level Discord summaries
Unsloth AI (Daniel Han) Discord
- Struggles with Unsloth Module Imports: Users encountered installation errors related to Python environments while importing the Unsloth module, with suggested fixes including using pip within a conda environment.
- This sparked a broader concern about dependency management which led to shared troubleshooting links and tips.
- Fine-Tuning Models is Tricky Business: Participants addressed difficulties in fine-tuning language models, noting models often failed to respond to queries post-training.
- Recommendations emphasized a careful evaluation of fine-tuning datasets to ensure optimal performance.
- WSL2 Recommended for Development: Windows users were advised to utilize WSL2 for running AI development environments effectively, including installing and executing models.
- Troubleshooting problems with WSL2 installations circulated among users, highlighting the need for guidance on specific errors.
- LLaMA 3 vs Claude 3.5 Sonnet Showdown: A user sought insights on how LLaMA 3 compares to Claude 3.5 Sonnet for coding tasks, hinting at a desire to enhance LLaMA performance with Unsloth.
- This interest indicated a larger conversation around adapting models for specific task effectiveness.
- Hugging Face Status Check: A user reported that services on Hugging Face were operational, despite their own troubles downloading models.
- This raised questions about potential localized issues versus broader accessibility.
HuggingFace Discord
- Hugging Face Service Experiencing Downtime: Users reported encountering server errors like 504 and 502 with Hugging Face services, indicating potential downtime. Community members shared their experiences and noted that services intermittently came back online.
- The ongoing issues seem to affect various functionalities, prompting discussions about server reliability and user frustrations.
- Sought Multilingual Embedding Models: Members discussed the need for recommendations on the best multilingual embedding models, especially for the German language. Emphasis was placed on selecting models suited for diverse linguistic applications.
- Various members chimed in on the importance of effective embedding models for high-dimensional spaces like multilingual datasets.
- Skepticism Surrounds Tesla Robot Event: Participants expressed doubts about the authenticity of the Tesla robot event, questioning whether the bots were actually operating autonomously. Many believed the robots might have been controlled remotely.
- Concerns about the implications for company reputation and investor perception highlighted the potential fallout from such misleading exhibitions.
- AI Agents and Collaboration Platform: A member introduced the idea of creating a customization platform for AI agents, debating the complexities faced by average users with existing solutions. The discussion quickly pivoted to the need for collaborative projects.
- Participants acknowledged an interest in more streamlined collaboration rather than scattered individual efforts.
- Clarifying Model Licensing Variants: Discussion surfaced regarding the distinctions between MIT and Apache licenses, focusing on aspects of commercial use and code forking. Members clarified that the MIT license is more permissive, appealing for versatile projects.
- The community expressed a preference for the flexibility afforded by MIT in various development scenarios.
LM Studio Discord
- Users voice LM Studio installation issues: Concerns arose about LM Studio installing without UAC prompts, especially regarding impacts on user profiles vs system files.
- Some users reported missing libraries when running appimages on certain Linux distributions, complicating setup.
- Qwen-2.5 leads in performance: Users compared performance across LLMs like Qwen-2.5 and Deepseek, noting Qwen's speed and efficiency for Python jobs.
- There was keen interest in testing various quantization options to further enhance output quality and speed.
- Scrutinizing GPU power management: Concerns emerged about low GPU utilization on dual 3060 cards running models at under 10%, despite achieving 17.60 tok/sec.
- Discussion hinted at potential IO bound challenges or erratic power management as culprits.
- NVIDIA holds edge for AI tasks: Debate centered on choosing between an NVIDIA 4060 Ti and an AMD RX 7700 XT, underscoring NVIDIA's superior AI support.
- Users suggested that NVIDIA GPUs generally lead to fewer complications in running AI applications.
- Mistral Large shines in consumer rigs: The Mistral-Large 123b model is prized for its flexibility on consumer-grade machines, particularly M2 Studio setups.
- Users noted that Mistral Large configurations efficiently utilize VRAM, handling various contexts adeptly.
Eleuther Discord
- OpenAI's Model Performance Concerns: Members expressed worries about OpenAI's model reportedly manipulating its testing environment, raising AI alignment issues.
- This highlights existing challenges in AI safety and ethics within the community.
- FA3 Slows Down Compared to F.sdpa: Users encountered significant challenges with FA3, noting it performed slower than F.sdpa, complicating the implementation process.
- One user highlighted confusion over the proper installation compared to existing models.
- NanoGPT Breaks Training Speed Records: A new NanoGPT speed record of 3.28 Fineweb validation loss in 15.2 minutes was achieved through code optimizations.
- Updates included using the SOAP optimizer and zero-initializing projection layers to enhance performance.
- Swiglu vs ReLU²: Activation Function Showdown: Discussion compared the effectiveness of ReLU² and Swiglu activation functions, suggesting different performances based on model size.
- Results indicated Swiglu may be more effective with larger models, although current tests favored ReLU².
- Creating a Hyperparameter Scaling Guide: A proposal for a guide on scaling hyperparameters emerged, aimed at centralizing knowledge for tuning methodologies crucial for model performance.
- Members acknowledged that existing information is largely held by researchers, making access difficult.
OpenAI Discord
- AI Aids Elderly Care Management: Participants discussed using AI to assist elderly individuals in managing medication and providing companionship, addressing reliability and ethical implications.
- Concerns were raised about ensuring that AI can handle care tasks without compromising safety.
- F5-TTS Voice Cloning Challenge: A user shared experiences with the F5-TTS model integrated with Groqcaster for automated voice outputs, which impressively manages local voice cloning.
- While quality isn't yet on par with ElevenLabs, the ability to generate everything locally is a significant perk.
- Spatial Computing Device Showdown: Users evaluated spatial computing devices like Meta Quest and Xreal for desktop use, debating their effectiveness in multi-monitor setups.
- While Meta Quest was favored for native application support, some limitations in optical quality were highlighted.
- GPT Integration Bug Report: Users noted that custom GPTs can no longer be integrated into another GPT's conversation with the '@' symbol, a shift that might indicate a bug.
- They suggested reaching out to support since some users still managed to use this function.
- Text2SQL Query Concerns: There's active discussion surrounding experiences with text2sql implementations, particularly in managing complex queries with LLMs.
- Users emphasized the need to keep context clear to avoid overwhelming outputs while fetching relevant data.
OpenRouter (Alex Atallah) Discord
- Inflection Models Go Live: The @inflectionAI model, powering @Pi, is now available on OpenRouter with no minimum spend and a playful focus on emojis 🍓.
- This model aims to enhance user interactions by allowing for a more engaging and fun chat experience with emoji integration 🤗.
- Grok 2 Launches with Exciting Access: OpenRouter now offers Grok 2 and Grok 2 Mini at a rate of $4.2/m input and $6.9/m output, though initially rate limited, as detailed in their announcement.
- Users appreciate the robust capabilities but note the critical nature of resource management during interactions.
- MythoMax Endpoint Offers Free Access: OpenRouter has launched a free MythoMax endpoint, broadening accessibility for users looking to leverage advanced models.
- This initiative aims to enhance user experience by providing more choices without additional costs.
- Chatroom Improvements Enhance Usability: Users can now drag and drop or paste images directly in chatrooms, enhancing overall interaction quality.
- These improvements reflect OpenRouter's commitment to streamlined and user-friendly communication within its platform.
- Grok API Running into Issues: Frequent errors like '500 Internal Server Error' and 'Rate limit exceeded' are reported by users encountering issues with the Grok API, which remains classified as experimental.
- It’s advised to consider beta models and other alternatives to mitigate these problems.
aider (Paul Gauthier) Discord
- Aider AI LLC Secures Source Code Home: The establishment of Aider AI LLC ensures that the aider source code is held under the Apache 2.0 license, maintaining its status as a completely free and open source project.
- “This is a community-driven effort with no funding rounds or employees involved,” reaffirming commitment to open-source principles.
- Users Overcome Aider Installation Challenges: Feedback indicated that installing Aider through
pipx
drastically simplified the setup process, avoiding lengthy install problems.- A user highlighted that adhering closely to the installation guide could mitigate installation issues.
- Jetbrains Plugin Crashes Spark Debate: Users reported that the Jetbrains plugin for Aider crashes on launch, pushing some to use Aider directly via terminal.
- Discussion focused on the plugin's lack of essential features, including file capture and keybindings, leading to frustration.
- Caution on Aider with Corporate Codebases: Concerns arose about using Aider on corporate codebases due to potential policy violations and risks of data leakage.
- While some emphasized that Aider operates locally without data sharing, worries about API use and screen sharing persisted.
- Comparative Performance of LLM Models: Debate over the effectiveness of various LLMs when integrated with Aider led to discussions about model performance, particularly regarding models like Grok-2 and GPT-4o.
- Members noted a need for careful selection of models to ensure optimal outputs in coding tasks.
Nous Research AI Discord
- Nous Research evolves into a startup: Nous Research, originally a Discord group, has transitioned into a funded startup focused on AI development, particularly in open-source projects.
- The community now plays a crucial role in facilitating collaboration and sharing ideas in the realm of AI research.
- DisTrO accelerates model training: DisTrO is designed to enable faster AI model training across the internet, promoting community-driven development as an alternative to closed models.
- The initiative aims to ensure sustained progress in the open-source field.
- Model collapse in neural networks revealed: A recent study examined the model collapse phenomenon, indicating that even 1% synthetic data can lead to significant performance decay.
- The research warns that larger models may exacerbate this collapse, challenging conventional scaling approaches.
- GSM-Symbolic improves LLM evaluations: The introduction of the GSM-Symbolic benchmark offers enhanced metrics for evaluating mathematical reasoning capabilities in LLMs.
- This benchmark diversifies assessment methods, promoting more reliable evaluations of language models.
- OpenAI faces scrutiny over Swarm framework: Accusations emerged alleging that OpenAI infringed on Kye Gomez's Swarms framework, with claims of stolen code and methodology.
- Potential legal actions are being considered unless investments are directed towards their project.
Perplexity AI Discord
- Reasoning Mode Launches in Pro Search: The Perplexity Team rolled out Reasoning Mode, an experimental feature that detects when extra compute can improve answers, encouraging users to share use cases in the feedback channel.
- Users provided various inquiries for Pro Search examples, including finding OpenAI co-founders and top-rated films, aiming to utilize the enhanced functionality.
- Navigating Cost Implications of AI Image Generation: Discussions surfaced regarding the costs tied to AI image generation, urging users to consider budgeting for these capabilities, more details found here.
- This dialogue highlighted the balance between affordability and project demands for high-quality visual output.
- User Frustrations with API Source URLs: Users grappled with the API not displaying source URLs in responses, prompting a call for support that met with silence, leaving inquiries unanswered.
- Discussion turned toward the online API, with references to
sonar-online
models available at Perplexity API Docs, aiming to clarify model functionalities.
- Discussion turned toward the online API, with references to
- Mixed Feedback on AI Model Performance: Users expressed mixed experiences using various AI models, with some favoring Claude for coding tasks over recent updates impacting O1 mini performance.
- Concerns were raised about the Perplexity API's ability to deliver quality responses similar to online interactions, highlighting significant discrepancies.
- Exciting Updates in Perplexity Pro Features: Updates in Perplexity Pro sparked recent interest as users shared insights on new features aimed at enhancing engagement and functionality.
- Members can explore these changes further via this link, fueling active discussions on best practices.
GPU MODE Discord
- Attention Layer Implementation Confusion: A member seeks a tutorial for an Attention layer using cuDNN’s SDPA in Python, feeling lost on instantiating pygraph. They follow a notebook from the cudnn-frontend repository.
- Any help would be appreciated in clarifying the implementation details.
- Performance Profiling Discrepancies in PyTorch and Triton: Members find significant differences in performance results while profiling with PyTorch, Triton, and CUDA, leading to questions on which profiler to trust.
- Despite Triton claiming equal performance overall, self-evaluations seem to place PyTorch ahead in many tests.
- Metal Programming Challenges on Apple Silicon: Members report difficulties in using Docker with Apple Silicon GPUs, citing unresolved issues within the community. An internal ticket for the problem remains open without active work.
- Discussion also touches on a PR for the torch.special.i0 operator, focusing on MPS support enhancements.
- Entropix Sampling Skepticism: Skepticism flares around Entropix sampling, with some alleging it feels like nonsensical cult stuff, raising questions about its credibility.
- Despite concerns, a recent blog post mentions its aim to simplify reasoning without extensive modifications.
- Strategies for Efficient LLM Deployment: An online meetup is scheduled on October 5th at 4 PM PST to discuss LLM deployment, featuring contributions from SGLang, FlashInfer, and MLC LLM.
- Topics include low CPU overhead scheduling and kernel generation for performant LLM serving, with opportunities for community interaction.
Cohere Discord
- CohereForAI Contributions Encouraged: Members emphasized the importance of meaningful contributions to the CohereForAI community, suggesting citizen science as an entry point for those engaged in AI.
- One individual expressed a desire to contribute and mentor, aligning projects with a vision for a symbiotic relationship with technology.
- AI Innovators Win Nobel Prize: Sir John J. Hopfield and Sir Geoffrey E. Hinton received the 2024 Nobel Prize in Physics for their groundbreaking contributions to AI and neural networks.
- Their work has laid the foundational discoveries crucial for advancing machine learning technologies.
- API Tokens Confusion: A member questioned the necessity of using
<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>
tokens for API requests, raising concerns about the quality of responses without them.- Will the responses still be decent without including these tokens? remains an unanswered question in the community.
- Clarification Needed on Cohere Rerank Pricing: There was confusion regarding whether web-search pricing is included in Cohere's Rerank pricing structure, as members couldn't find relevant details on the site.
- Understanding this pricing model is vital for planning effective implementation strategies.
- Upcoming Gen AI Hackathon Announcement: Members are invited to participate in the Gen AI Hackathon organized by CreatorsCorner, aiming to create innovative multi-agent systems.
- This event encourages collaboration to enhance human potential through intelligent solutions, as noted in the hackathon invite.
Modular (Mojo 🔥) Discord
- Mojo Installation Frustrations: Users are encountering installation issues with Mojo, leading to broken playground functionality and demo code errors, with a lack of clear tutorials available.
- One user noted that this disjointed process for Magic, Mojo, and MAX is causing significant confusion in the community.
- AES Hardware Support Advances: A member showcased their progress on implementing AES hardware support in Mojo via LLVM intrinsics, enhancing library integration capabilities.
- This effort reinforces Mojo's flexibility, allowing for advanced hardware functionalities to be incorporated smoothly into projects.
- Compilation Times Under Review for MAX: Compilation times for MAX are around 300 ms for graphs post-initial run and 500 ms for the first compilation, even for simple tasks.
- Discussion highlighted the importance of improving cache hit times to optimize performance during development.
- Implicit Conversions Create Debate: Implicit conversions in Mojo Lists raised questions among members, as adding an Int to a List seems possibly unintended due to existing constructors.
- An ongoing issue is tracking this behavior, which might complicate type handling in future implementations.
- Building with Mojo Faces Linking Errors: Users faced linking errors while attempting to build Mojo files, indicating potential missing libraries during the compilation process.
- Assistance included checking magic environment activation and proper installation protocol via command line.
Latent Space Discord
- OpenAI Swarm Launches Experimental Framework: OpenAI introduced Swarm, a lightweight library for building multi-agent systems, highlighting a stateless abstraction for managing agent interactions without relying on the Assistants API.
- It’s an experimental framework, aiming to facilitate easy understanding of agent roles and handoffs.
- Entropix Gains Popularity Among Members: Members engaged in an enthusiastic discussion about Entropix, providing an overview that shed light on its functionality and potential impacts.
- As interest grows, users are eager to see forthcoming evaluation features linked to the tool’s progression.
- RAG Techniques for Enhanced AI Performance: Discussion on RAG Techniques centered around a GitHub repository showcasing advanced methods for integrating retrieval with generative models.
- Participants aim to optimize performance, comparing frameworks like Haystack to custom solutions for specific use cases.
- Insights from Jensen on NVIDIA's Infrastructure: In a recent interview, Jensen discussed NVIDIA's full stack approach to AI infrastructure, underlining the imperative for accelerated computing in modern applications.
- His remarks reaffirmed generative AI’s transformative potential and indicated the continuous need for innovation in this space.
- Production AI Engineering Episode Highlights: The latest podcast episode covered insights into Production AI Engineering, focusing on the critical role of Evals in the industry.
- Experts declared Evals central to the landscape of LLM Ops, emphasizing the need for robust evaluation metrics as a growing priority.
Stability.ai (Stable Diffusion) Discord
- 3060ti Proves Competent in Stable Diffusion: Discussion highlighted the effectiveness of the 3060ti for Stable Diffusion, performing surprisingly well despite its 8GB VRAM limitations. Users cited Flux image generation as a testament to the GPU's capabilities.
- One user asserted that, with the right techniques, the 3060ti can handle demanding tasks in AI image generation efficiently.
- Lora Training Outshines Embedding: Participants debated the benefits of Lora training over embedding, asserting that Lora typically results in higher quality images. While embedding impacts only the text encoder, Lora allows for more nuanced diffusion model training.
- This detail sparked interest in deeper discussions about workflow adjustments for optimal image quality.
- Image Upscaling Techniques Under Scrutiny: The community compared Tiled Diffusion with Ultimate SD Upscale, noting each method serves distinct purposes—VRAM management vs. resolution enhancement. The merits of both techniques were extensively evaluated in ongoing projects.
- Users agreed that understanding when to apply each technique can significantly affect the outcome of image processing tasks.
- Image to 3D Model Generation Still Needs Work: The complexities of image to 3D model generation generated considerable discussion, as participants recognized the existing gaps in effective solutions. Multi-view inference techniques emerged as the most reliable methods currently available.
- Members expressed a collective need for innovation in this area, as the challenges remain significant.
- Looking for Help with Product Photo Integration: A member sought advice on integrating product photos into various backgrounds, emphasizing the need for high-quality results over basic compositing. Suggestions pointed towards leveraging Lora training to achieve better blending in final images.
- The conversation underscored the importance of advanced techniques in fulfilling specific visual demands.
LlamaIndex Discord
- LlamaIndex Hackathon Approaches: The upcoming LlamaIndex Hackathon this weekend invites participants to engage and innovate, with all relevant slides and resources shared for pre-hack preparation.
- Participants are encouraged to utilize these resources to ensure their projects are well-grounded.
- Building RAG Pipelines Simplified: Check out this video that walks through a basic RAG pipeline setup using LlamaIndex, highlighting essential workflows and components.
- The tutorial features a simple implementation with the router query technique for improved accuracy.
- Integrating Chat History with RouterQueryEngine: An inquiry about the RouterQueryEngine indicated an interest in incorporating chat history through the inclusion of all chat messages to enhance interaction dynamics.
- Workflow suggestions and examples were shared to facilitate better integration practices.
- Challenges with PDF Image Extraction: Users faced difficulties with extracting images from PDFs, frequently encountering unexpected ASCII characters in outputs, creating confusion.
- Guidance was sought to clarify data exports from parsed results, indicating a need for better documentation or support.
- Interest in Colpali within LlamaIndex: Questions arose about the potential for Colpali implementation in LlamaIndex, amid a noted gap in documentation.
- While full embedding isn't currently supported, community interest suggests that adding it as a reranker could be on the horizon.
tinygrad (George Hotz) Discord
- Tinygrad Type Annotations Discussion: The team evaluated three PRs for adding type annotations in Tinygrad, dismissing one due to performance concerns and another recognized contributor prioritized higher.
- A PR was rejected after failing tests, prompting worries about the practicality of merging such changes.
- Bounties Require Proven Contributors: George emphasized that contributors with several merged PRs are prioritized for bounty tasks, noting a new $200 bounty for parallel SHA3 implementation.
- This highlights the necessity for experience as a pre-requisite for tackling larger contributions.
- Challenges in SHA256 Implementation: A proposal for a complete SHA256 implementation in Tinygrad sparked talks about integrating parallel processing despite current design limitations.
- George showed interest in exploring parallel capabilities to optimize the implementation.
- DDPM Schedulers Thrive on Metal: A member introduced their own DDPM scheduler for training diffusion models on Metal, filling a gap in Tinygrad's resources.
- They are willing to collaborate with others needing support with this new tool.
- Addressing Tensor Gradient Axis Issues: The community debated solutions for resolving gradient axis mismatches in tensors, offering multiple approaches like axis alignment and resharding.
- Concerns were raised about the wastefulness of resharding as a solution.
Interconnects (Nathan Lambert) Discord
- OpenAI O1 Model Replication Advances: The initial report on replicating OpenAI's O1 model showcases a new 'journey learning' paradigm that enhances mathematical reasoning with 327 training samples, resulting in an 8% improvement in performance.
- Microsoft's Top AI Researcher Joins OpenAI: Reports confirm that Sebastien Bubeck, a prominent AI researcher from Microsoft, is moving to OpenAI, raising questions about the motivations behind such transitions amid lucrative AI roles. The Information highlights this significant career shift.
- The movement has created a stir, with industry colleagues humorously speculating on the implications for existing AI teams.
- Ex-OpenAI Employees Launching Startups: A staggering 1,700 startups are anticipated to be founded by former OpenAI employees, marking a significant surge in the AI startup ecosystem.
- This trend reflects a shift toward innovation and diversification within the field, producing potential new leaders in AI technology.
- Dario Amodei's Influential Work Gains Recognition: Machines of Loving Grace has been lauded for its compelling title and engaging content, stirring interest in AI's potential benefits for society.
- This growing discourse signals a shift towards positive perceptions of AI's future, moving away from fear-based narratives.
- Folding@Home's Early Influence in AI: Discussion arose around Folding@Home and its perceived underwhelming impact, with some members asserting it was ahead of its time despite its pioneering contributions to biological computing.
- The conversation also acknowledged the relevance of established methods like docking in drug discovery that seemed overshadowed during the Nobel discussions.
DSPy Discord
- Next.JS Voice Interview Prep Platform Launch: A member announced the development of a full stack Next.JS voice interview prep/testing platform aimed at enhancing interview preparation through voice interaction.
- This platform is expected to significantly improve the user experience during interview training.
- GraphIC Transforms ICL Selection: The paper introduces GraphIC, a technique using graph-based representations and Bayesian Networks to enhance the selection of in-context examples for LLMs.
- It resolves limitations of text-based embedding methods in multi-step reasoning tasks by filtering out shallow semantics.
- LLM Classifier Seeks Ambiguity Handling: A user is working on training an LLM classifier and is seeking community input on handling classification ambiguities to effectively manage uncertain outputs.
- The suggestion involves adding a second output field in the LLM signature to declare ambiguities instead of creating separate classes.
- Assessing Output Effectiveness with Cosine Similarity: A member inquired about a metric to evaluate if chatbot outputs meet established criteria, considering cosine similarity to compare input queries with generated categories.
- Stuart is actively seeking suggestions to refine this approach for better off-topic detection.
- FastAPI Route Creation for Signatures: A member shared a code snippet that enables any dspy.Signature to turn into a FastAPI route, returning the predictor as a dictionary with init_instant function used for environment initialization.
- This implementation streamlines the request processing essential for developing APIs with DSPy.
LAION Discord
- LLaMA 3.2 Dominates Pop Culture Knowledge: LLaMA 3.2 excels over competitors with its training on 5 billion images, allowing for more coherent captions.
- In comparisons, LLaMA 3.2 demonstrates significant contextual understanding compared to models like Molmo and PixTral.
- PixTral Shines in Adult Content Scenarios: Members highlighted that PixTral stands out when focused on adult content, unlike LLaMA 3.2, which is better suited for broader contexts.
- The contrast indicates that while PixTral has a niche, LLaMA 3.2 maintains cultural relevance across more general applications.
- Epic Games Removing Sketchfab Sparks Concerns: Epic Games' removal of Sketchfab is set to eliminate 800k 3D models from Objaverse, prompting users to download urgently.
- This decision has raised alarms about its effects on the 3D modeling community and users relying on those resources.
- o1-mini Can't Compete with o1-preview: Reports indicate that o1-mini is outperformed by o1-preview, being described as brittle on straightforward tasks according to recent insights.
- Despite earlier claims of matching larger models, evidence suggests o1-preview excels even on Olympiad level tasks.
- Challenges with CLIP's Contrastive Training: Using CLIP for training T2I models speeds up the process but introduces artifacts tied to its contrastive training methods.
- These artifacts raise concerns about impacting overall training quality, suggesting trade-offs in efficiency and performance.
LLM Agents (Berkeley MOOC) Discord
- Graham Neubig's Lecture on AI Agents: Graham Neubig's lecture today at 3:00 PM PST discusses
- Neubig also highlights complexities in developing AI agents for large repositories, addressing issues such as file selection and integration of web browsing into workflows.
- Registration Delays and Troubles: Members confirmed that course registration remains open until Dec 12, with a successful sign-up experience shared after troubleshooting a linking issue.
- Participants reported challenges with Google Forms for quizzes, suggesting clearing browser cache resolved access issues; course timings are set for 3:00 PM to 5:00 PM PST.
- Defining the AI Agent: An AI agent autonomously performs tasks via interactions with APIs and databases, with ChatGPT classified as an agent, unlike gpt-3.5-turbo, which lacks such capabilities.
- Discussion also includes ongoing efforts to refine AI agent definitions, emphasizing the importance of community input via platforms like Twitter.
- Chain of Thought Enhances LLM Problem Solving: The Chain of Thought (CoT) methodology assists LLMs in breaking down complex tasks into manageable steps, promoting clarity in problem-solving.
- Members recognized CoT’s effectiveness through an example involving Apple, showcasing how systematic breakdowns lead to final solutions.
- AI-Powered Search Book Gains Attention: A member recommended this book as the go-to resource for AI-powered search, praising its anticipated impact over the coming years.
- The book is expected to serve as a vital reference for both AI practitioners and researchers, highlighting its future relevance in the field.
Torchtune Discord
- Gemma 2 Support leverages Flex Attention: Discussion focused on implementing Gemma 2 using Flex Attention, with logit softcapping identified as the main blocker that needs a proper
score_mod
function.- Members believe the tradeoff for Flex simplifies the process, although it may require high compute capabilities with CUDA.
- Introduction of the Aria Model: Aria, a new open multimodal AI model, showcases a 3.9B and 3.5B parameter architecture and excels in language and coding tasks, outperforming Pixtral-12B.
- While there are no direct benchmark comparisons yet, early indications show Aria’s capabilities surpass its contemporaries.
- LegoScale Revolutionizes Distributed Training: LegoScale introduces a customizable, PyTorch-native system for 3D parallel pre-training of large language models, significantly boosting performance.
- Its modular approach aims to simplify complex training across GPUs, potentially changing the landscape of distributed training.
- Insights from the State of AI 2024 Report: The State of AI Report 2024 by Nathan Benaich outlines significant trends and investment areas in AI with few mentions of models like Torchtune.
- This report serves to prompt discussions on the future of AI, particularly regarding its applications in medicine and biology.
- Out-of-Shared-Memory Issues with Flex Attention: A GitHub issue shared problems with flex attention on the RTX 4090, detailing errors linked to out-of-shared-memory problems.
- The conversation included a minimal reproduction code snippet, fostering collaboration for troubleshooting.
LangChain AI Discord
- Swarm.js Launch for Node.js Enthusiasts: Swarm.js, a lightweight Node.js SDK, orchestrates multi-agent systems using the OpenAI API, enabling seamless agent management and task execution.
- Developers can easily start by running
npm install openai-swarm-node
, with the project actively inviting contributions and collaboration from both beginners and experts.
- Developers can easily start by running
- Community Closure Announcement: Jess announced that the LangChain Discord community is set to close on October 31, 2024, to focus on creating a new community platform.
- All members are encouraged to fill out a form for updates and provide feedback at community@langchain.dev, with invitations extended for potential moderators.
- Exploring Contextual Retrieval Techniques: A new YouTube video illustrates how to implement contextual retrieval using LangChain and OpenAI’s Swarm Agent, guiding viewers through the integration process.
- This informative content is aimed at enhancing information retrieval, making it especially relevant for those employing LangChain in their projects.
- bootstrap-rag v0.0.9 Goes Live!: bootstrap-rag v0.0.9 has been released with critical bug fixes, improved documentation, and integration with LangChain and MLflow-evals.
- The update also includes templates for Qdrant, enhancing retrieval-augmented generation capabilities, a key area for AI engineers focusing on efficient data handling.
- LangGraph Tutorial for Job Seekers: A new tutorial demonstrates building a two-node LangGraph app that analyzes resumes against job descriptions, offering practical benefits for job applicants. Watch here.
- The app can craft tailored cover letters and generate job-specific interview questions, making it a handy tool for those new to LangGraph.
OpenAccess AI Collective (axolotl) Discord
- Exploring Instruction Data Impact: A member raised a question about the use of instruction data during pretraining, emphasizing the potential benefits for model engagement.
- This topic could lead to a discussion on innovative pretraining techniques that enhance model adaptability.
- Config Sharing Sparks Suggestions: In a discussion on config sharing, a member requested a specific config while suggesting sample packing as a crucial update.
- Challenges with multi-GPU setups were highlighted, emphasizing the need for a thorough setup review.
- Fine-Tuning with Adapters: There was discussion on merging an existing Llama 3 adapter with a fine-tuned model to improve accuracy in tasks.
- A GitHub guide was shared for the merging process, reiterating the importance of proper config settings.
- Text Completion Training Enhances Instruction Models: Training instruct models like GPT-3.5-Instruct on text completion tasks can lead to marked performance improvements in instruction compliance.
- Community members cautioned about overfitting risks, advising diverse datasets for optimal training outcomes.
- Diversity Key to Avoiding Overfitting: Concerns about overfitting emerged when discussing training datasets, with a call for more diversity to enhance generalization.
- Members stressed monitoring performance across tasks to mitigate risks of degradation on unfamiliar datasets.
OpenInterpreter Discord
- Message Format Compliance Reminder: Members were reminded to adhere to the prescribed message format in the channel to maintain organization and clarity.
- The reminder emphasized the importance of following established guidelines to enhance channel communication.
- Reference to Channel Rules: A reference was made to existing rules guiding behavior and contributions within discussions in the channel.
- Members were encouraged to review these rules for better channel dynamics.
- Aria Becomes the Top Multimodal Model: Aria, by @rhymes_ai_, is now ranked first on the 🤗 Open LLM Leaderboard, showcasing 24.9B parameters and handling image, video, and text inputs.
- Users Praise Aria's Multimodal Capabilities: Users are enthusiastic about the 25.3B multimodal Aria model, calling it the BEST vision language model I have ever tried!
- Released under the Apache-2.0 license, fine-tuning scripts are also available for community engagement.
- AI Reasoning Capabilities Debated: A YouTube video titled 'Apple DROPS AI BOMBSHELL: LLMS CANNOT Reason' raises critical questions about language models' reasoning capabilities.
- The creator stimulates a dialogue around current AI limitations, urging viewers to prepare for AGI advancements.
AI21 Labs (Jamba) Discord
- Member Starts Support Inquiry on Jamba: A user initiated a support thread about Jamba issues, asking if the channel was appropriate for assistance.
- Another member confirmed they had addressed the inquiry in the same channel, promoting ongoing discussions there.
- Importance of Thread Continuity for Jamba Issues: In response to the support inquiry, a member emphasized keeping the discussion within the original thread to maintain clarity.
- They pointed out that this approach would facilitate easier access to pertinent information in the future.
Mozilla AI Discord
- Panel Discussion on Community Engagement Strategies: A panel of Community and Developer Relations experts, including Jillian Bejtlich and Rynn Mancuso, is set to discuss actionable strategies for enhancing community engagement during the upcoming event.
- This engagement-focused session aims to equip attendees with practical techniques for growing user bases and boosting contributions to projects.
- Tactical Insights for Building Thriving Communities: The panel will share tactical advice on how to cultivate a successful community around projects, emphasizing the importance of relationship-building beyond coding.
- Project leads seeking to refine their community-building skills will find this session particularly useful for networking and strategy enhancement.
- RSVP for the Community Panel Discussion!: Participants are encouraged to RSVP for the panel discussion on community-building practices here to secure their spot.
- “Don’t miss out on this invaluable opportunity!” resonates within the channel, urging engagement from the community.
DiscoResearch Discord
- Introducing the Backtrack Sampler: Mihai4256 shared an interesting GitHub repository focused on a backtracking sampling method.
- This repository is likely to attract AI engineers interested in advanced sampling techniques and model optimization.
- Check Out the GitHub Repo: The repository provides innovative approaches to sampling that could enhance algorithm efficiency and model accuracy.
- Mihai4256 encourages collaboration and feedback from the community on the implementation discussed in the repo.
Gorilla LLM (Berkeley Function Calling) Discord
- Performance Struggles in Multi-Turn Evaluations: Members encountered a ~0% performance rate during multi-turn evaluations, as the model fails to exit the count loop despite correct predictions.
- The discussion highlighted various efforts to find a viable solutions for this evaluation conundrum.
- Workaround Boosts Multi-Turn Evaluation Ratings: A temporary code modification in base_handler.py improved evaluation accuracy to ~15% performance by attempting each round only once.
- However, the need for compliance with modification restrictions has left members seeking alternative strategies to enhance performance.
The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The LLM Finetuning (Hamel + Dan) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
PART 2: Detailed by-Channel summaries and links
Unsloth AI (Daniel Han) ▷ #general (375 messages🔥🔥):
Importing Unsloth module
Fine-tuning models for specific tasks
Using WSL2 for development
Training procedures and options
Loading local model paths
- Issues with Importing Unsloth Module: Users faced challenges importing the Unsloth module, often encountering installation errors related to Python environments and dependencies.
- A recommended solution involves installing Unsloth through pip within a conda environment.
- Challenges in Fine-tuning Models: Participants discussed difficulties in fine-tuning language models, particularly when models failed to respond to basic queries after training.
- The community suggested evaluating the fine-tuning dataset and model type to ensure effectiveness.
- Using WSL2 for Development: Windows users were advised to run their AI development environments in WSL2 to properly install and run models like Unsloth.
- Some users experienced installation issues, prompting others to share troubleshooting tips and links for resolving WSL2 errors.
- Training Procedures and Options: Discussions included various training scripts for models, with participants clarifying steps for running them locally after successfully setting up their environments.
- Users confirmed that the training scripts in notebooks matched those needed for local execution.
- Loading Local Model Paths: Concerns arose about how to specify local paths for already downloaded models, with users asking for guidance on correctly referencing these paths.
- The community emphasized using proper syntax to ensure models load correctly in the training scripts.
- Introducing SuperNova-Medius: Arcee AI's 14B Small Language Model That Rivals a 70B: First came our flagship 70B SuperNova, followed by the 8B SuperNova-Lite. Today we add to this family of superpower Small Language Models with the release of the 14B SuperNova-Medius.
- Google Colab: no description found
- arcee-ai/SuperNova-Medius · Hugging Face: no description found
- unsloth/Llama-3.2-11B-Vision · Hugging Face: no description found
- Google Colab: no description found
- unsloth/Hermes-3-Llama-3.1-8B · Hugging Face: no description found
- NousResearch/Hermes-3-Llama-3.1-8B · Hugging Face: no description found
- Reddit - Dive into anything: no description found
- Sorry Im Sorry GIF - Sorry Im Sorry Sad - Discover & Share GIFs: Click to view the GIF
- GitHub - arcee-ai/DistillKit at blog.arcee.ai: An Open Source Toolkit For LLM Distillation. Contribute to arcee-ai/DistillKit development by creating an account on GitHub.
- output_sample: no description found
- Kanna Kamui Miss Kobayashi GIF - Kanna Kamui Miss Kobayashi Dragon Maid - Discover & Share GIFs: Click to view the GIF
- unclemusclez/unsloth-qwen2.5-coder: Qwen 2.5 Coder with Unsloth
- Unproprietary Corporation: Commercial Open Source Software (COSS) infrastructure platform for collaborative work - Unproprietary Corporation
- GitHub - microsoft/VPTQ: VPTQ, A Flexible and Extreme low-bit quantization algorithm: VPTQ, A Flexible and Extreme low-bit quantization algorithm - microsoft/VPTQ
- Kiss GIF - Kiss - Discover & Share GIFs: Click to view the GIF
- GitHub - unslothai/unsloth: Finetune Llama 3.2, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory: Finetune Llama 3.2, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory - unslothai/unsloth
- Package repository for pytorch | Anaconda.org: no description found
- Package repository for nvidia | Anaconda.org: no description found
- Package repository for xformers | Anaconda.org: no description found
Unsloth AI (Daniel Han) ▷ #off-topic (6 messages):
LLaMA vs Claude 3.5 Sonnet
Training Models Issues
Hugging Face Status
- Comparing LLaMA 3 and Claude 3.5 Sonnet: A user inquired about how LLaMA 3 stacks up against Claude 3.5 Sonnet specifically for coding tasks.
- They expressed interest in tuning LLaMA with Unsloth to see its effectiveness.
- Nan Issues during Model Training: One user reported experiencing issues where their models suddenly generated NaN errors, while other aspects of training proceeded smoothly.
- This raises questions about potential underlying causes affecting model stability.
- Hugging Face Status Reported: A user shared a status update from Hugging Face, indicating that all services are currently online as of the last update.
- Despite the site being operational, another user mentioned they were unable to download any models.
Link mentioned: Hugging Face status : no description found
Unsloth AI (Daniel Han) ▷ #help (104 messages🔥🔥):
Finetuning Qwen 2.5
Using Llama 3.2 and Ollama for embeddings
Issues with GGUF conversion
Embedding comparison in RAG frameworks
VLLM usage for quantized models
- Finetuning Qwen 2.5 Model: A user encountered problems while trying to finetune the Qwen 2.5 0.5B model, specifically with local datasets in Alpaca format.
- They received advice to ensure proper dataset formatting and mapping keys, referencing Unsloth documentation for guidance.
- Llama 3.2 for Embeddings: A user using Ollama with Llama 3.2 3B encountered unexpected results when comparing animal trait embeddings, with rat consistently scoring highest.
- Despite detailed descriptions, the similarity scores did not align with expectations, prompting queries about RAG frameworks and embedding methods.
- Challenges in GGUF Conversion: A user confirmed issues with GGUF conversion for llama-3.1-70b-instruct models, as generated text was nonsensical after conversion while working fine without it.
- They shared code snippets used during the conversion process, seeking further advice on how to resolve the output problem.
- Using VLLM with Unsloth Models: A user experienced issues when trying to run unsloth/Qwen2.5-14B-bnb-4bit with VLLM, receiving an error regarding BitsAndBytes quantization.
- Advice was provided to use the non-quantized version of the model as the error indicated unsupported quantization.
- Embedding Comparisons in RAG Frameworks: A user inquired how RAG frameworks perform embedding comparisons, questioning the consistency of results across prompts.
- They shared sample results demonstrating that unrelated traits often produced anomalous similarity scores, indicating a need for further investigation.
- Chat Templates | Unsloth Documentation: no description found
- GitHub - unslothai/unsloth: Finetune Llama 3.2, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory: Finetune Llama 3.2, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory - unslothai/unsloth
Unsloth AI (Daniel Han) ▷ #research (13 messages🔥):
Adapter Training Techniques
LoRA Fine-tuning
Optimizing RNNs for Hardware
Training LLMs for Code Generation
Dataset Importance in Model Training
- Exploring Adapter Training before Output Layer: A discussion arose about training a small adapter before the output layer to alter embeddings towards Chain of Thought (CoT) reasoning.
- Concerns about whether this concept aligns with LoRA fine-tuning were raised, leading to clarification on their functional distinctions.
- LoRA Fine-tuning Attributes Discussed: Members explained that LoRA involves attaching adapters to existing layers, freezing the original model layers to train just the adapters.
- The balance of retaining original model knowledge while allowing for targeted training was emphasized, captured in the phrase, 'LoRA learns less forgets less.'
- RNN Optimization Strategy Revealed: A paper was shared discussing optimization of traditional RNNs to achieve performance levels akin to transformers using hardware advancements.
- The authors, through FlashRNN and Triton, demonstrated parallel processing capabilities while retaining state-tracking for tasks like logical reasoning.
- Training LLMs for Code Generation Inquiry: A question was posed regarding how to train an LLM for generating UI code, akin to services like V0.
- The suggestion included training on a specific repository with text-completion models, and potential challenges with fine-tuning without Q/A pairs were noted.
- Dataset Relevance in Model Training: The conversation highlighted the significance of the dataset in model training, stating that 'it's all in the dataset.'
- It was noted that understanding dataset construction aligns with fundamental data science principles and is supported by available courses and literature.
Link mentioned: FlashRNN: I/O-Aware Optimization of Traditional RNNs on modern...: While Transformers and other sequence-parallelizable neural network architectures seem like the current state of the art in sequence modeling, they specifically lack state-tracking capabilities....
HuggingFace ▷ #general (365 messages🔥🔥):
Hugging Face Service Status
Multilingual Embedding Models
AI Agent Discussions
Tesla Robot Event Skepticism
Model Licensing Explained
- Hugging Face Service Status Issues: Users reported experiencing various server errors, including 504 and 502 status codes, indicating potential downtime or performance issues with Hugging Face services.
- The community expressed frustration but also shared updates, with some confirming that services were intermittently coming back online.
- Best Multilingual Embedding Models: A user sought recommendations for the best current multilingual embedding models, specifically for the German language.
- Discussion points included the importance of finding a well-suited model for multilingual applications.
- Skepticism Around Tesla's Robot Event: Participants discussed skepticism regarding the authenticity of the Tesla robot event, highlighting suspicions that robots were controlled remotely rather than functioning autonomously.
- This led to broader conversations about the implications of such demonstrations for company reputation and investor perception.
- AI Agent and Collaboration Discussions: A user proposed the idea of creating a platform for customizing AI agents, expressing concerns about existing solutions being too complex for average users.
- This led to a conversation about the community's focus on individual projects rather than collaboration on overarching platforms.
- Understanding Model Licensing: Discussion emerged about the differences between MIT and Apache licenses, particularly in relation to commercial use and code forking.
- Users emphasized the permissiveness of the MIT license compared to Apache, highlighting its suitability for various projects.
- Hugging Face – The AI community building the future.: no description found
- Upscale Model Examples: Examples of ComfyUI workflows
- Fast Subtitle Maker - a Hugging Face Space by Nick088: no description found
- CompVis/stable-diffusion-v1-4 · Hugging Face: no description found
- com (Full name): no description found
- HuggingChat: Making the community's best AI chat models available to everyone.
- Theodd1sout Vsauce GIF - Theodd1sout Vsauce Or is it - Discover & Share GIFs: Click to view the GIF
- no title found: no description found
- Mlp Sunrise GIF - Mlp Sunrise Morning - Discover & Share GIFs: Click to view the GIF
- GitHub: Let’s build from here: GitHub is where over 100 million developers shape the future of software, together. Contribute to the open source community, manage your Git repositories, review code like a pro, track bugs and fea...
- - YouTube: no description found
- no title found: no description found
- Tweet from Troy Teslike (@TroyTeslike): It seems some people are still confused about how autonomous the Optimus bots were at the Tesla event. This clip is from another video Tesla released a while ago. It shows how the bots are controlled ...
- Google Colab: no description found
- catsOfpeople/owlaHightQ · Datasets at Hugging Face: no description found
- Hugging Face status : no description found
- advice & tips on indexing using multimodal embeddings 24h of screen recording on consumer hardware (100,000 frames/day) · huggingface/candle · Discussion #2557: hi, i'm working on: https://github.com/mediar-ai/screenpipe it records your screens & mics 24/7 and extract OCR & STT into a local sqlite db we want to explore vector search to improve sea...
- Search | arXiv e-print repository: no description found
- Throw Up Dry Heave GIF - Throw Up Dry Heave Vomit - Discover & Share GIFs: Click to view the GIF
- Smile Mlp GIF - Smile MLP Pinkie - Discover & Share GIFs: Click to view the GIF
- Mlp My Little Pony Friendship Is Magic GIF - Mlp My Little Pony Friendship Is Magic My Little Pony - Discover & Share GIFs: Click to view the GIF
- rombodawg (rombo dawg): no description found
- rombodawg/Everything_Instruct · Datasets at Hugging Face: no description found
HuggingFace ▷ #today-im-learning (8 messages🔥):
Neural Network in Rust
Collaboration on NN Project
Terraria Neural Networks
Incomplete NN Development
- Interest in Rust Neural Network Project: A member expressed interest in collaborating to build a neural network in Rust from scratch, although the project is noted as still incomplete.
- They clarified that they were referring to the programming language, not the video game Terraria.
- Collaboration Suggestion for NN Development: Another member suggested making a collaboration post to gather more interest in the neural network project.
- They provided a link to the relevant channel for organizing this collaboration: Collaboration Post.
- Discussion on Neural Networks in Video Games: Members jokingly discussed whether the request was for the programming language or the game, hinting at the creativity involved in building neural networks in Terraria.
- The conversation underscored the playful interactions present within the community surrounding neural network developments.
- Tweet from Ethan He (@EthanHe_42): I'm excited to share our latest research on improving LLM by upcycling them into Mixture of Experts (MoE)! 1. We upcycled the Nemotron-4 15B model on 1T tokens and compared it to a continuously tr...
- Upcycling Large Language Models into Mixture of Experts: Upcycling pre-trained dense language models into sparse mixture-of-experts (MoE) models is an efficient approach to increase the model capacity of already trained models. However, optimal techniques f...
- Megatron-LM/megatron/core/transformer/moe at main · NVIDIA/Megatron-LM: Ongoing research training transformer models at scale - NVIDIA/Megatron-LM
HuggingFace ▷ #cool-finds (8 messages🔥):
DIAMOND Agent on GitHub
Decentralized Training with INTELLECT-1
Red Bull Basement Innovation Event
Medical AI Research Highlights
OSS-Fuzz Gen Project
- Explore the DIAMOND Agent Architecture: The DIAMOND repository features a reinforcement learning agent trained within a diffusion world model, showcased at NeurIPS 2024.
- Its unique architecture promises advancements in learning environments, making it worth a look for AI researchers.
- INTELLECT-1 Decentralized Training Launch: INTELLECT-1 is announced as the first decentralized training for a 10B model, set to scale efforts significantly.
- This initiative invites contributors to join the movement towards developing open-source AGI, broadening community participation.
- Red Bull Basement 2024 Call for Innovators: The Red Bull Basement event encourages innovators to pitch their ideas with the potential of receiving AI-backed support.
- Teams can create business plans for a chance to compete at a World Final in Tokyo; applications are open until October 27, 2024.
- Weekly Insights into Medical AI: The latest podcast discusses top research papers in medical AI, featuring notable models like MMedAgent, focused on multi-modal agents for medical tool usage.
- You can catch the recap of essential models and methodologies from the week by visiting this YouTube video.
- OSS-Fuzz Gen for LLM Fuzzing: OSS-Fuzz Gen by Google introduces a new approach to fuzzing using LLM-powered techniques.
- This repository aims to enhance software robustness and invites collaboration for further development.
- Adaptability protein dynamics - a Hugging Face Space by MISATO-dataset: no description found
- Tweet from Prime Intellect (@PrimeIntellect): Announcing INTELLECT-1: the first-ever decentralized training of a 10B model Scaling decentralized training 10x beyond prior efforts. Anyone can join us to build open-source AGI 🦋
- - YouTube: no description found
- GitHub - 0xD4rky/Vision-Transformers: This repo has all the basic things you'll need in-order to understand complete vision transformer architecture and its various implementations.: This repo has all the basic things you'll need in-order to understand complete vision transformer architecture and its various implementations. - 0xD4rky/Vision-Transformers
- Calling all innovators: Red Bull Basement is back: A single good idea will launch innovators from the U.S. to the global stage as Red Bull Basement returns for 2024.
- GitHub - eloialonso/diamond at csgo: DIAMOND (DIffusion As a Model Of eNvironment Dreams) is a reinforcement learning agent trained in a diffusion world model. NeurIPS 2024 Spotlight. - GitHub - eloialonso/diamond at csgo
- - YouTube: no description found
- @aaditya on Hugging Face: "Last Week in Medical AI: Top Research Papers/Models 🏅 (October 5 - October…": no description found
- Tweet from Open Life Science AI (@OpenlifesciAI): Last Week in Medical AI: Top Research Papers/Models 🏅 (October 5 - October 12, 2024) 🏅 Medical AI Paper of the Week: MMedAgent: Learning to Use Medical Tools with Multi-modal Agent Authors: (@Haoy...
- GitHub - google/oss-fuzz-gen: LLM powered fuzzing via OSS-Fuzz.: LLM powered fuzzing via OSS-Fuzz. Contribute to google/oss-fuzz-gen development by creating an account on GitHub.
HuggingFace ▷ #i-made-this (7 messages):
OpenAI Swarm Framework
GitHub Token Markdown Generator
Qompass Diver Editor
Backtrack Sampler for LLMs
Vintage Action AI Video
- OpenAI Swarm Adds Support for Open Source Models: Support for open source models was added to the Swarm framework managed by the OpenAI Solutions team.
- A community member suggested checking a relevant issue on GitHub for improvements before submitting a pull request.
- Generate Markdown Structure for GitHub Repos: An easy tool allows users to enter their GitHub token and repository URL to generate a comprehensive Markdown file structure for seamless copying.
- This is available on GitHub as a React application aiming to enhance development workflows.
- Introducing the Qompass Diver Editor: The Qompass Diver is a fast and efficient editor currently in open preview, tested on various platforms including Arch Linux, Android, and Mac M1.
- It operates under a new AGPL/Qompass Dual-License, promoting an open-source ethos.
- Backtrack Sampler Framework for LLMs: A new framework for LLM sampling algorithms allows for the discarding of generated tokens, enabling users to regenerate incorrectly generated tokens.
- It includes demo algorithms and supports both GGUF models and Huggingface format models, available on GitHub.
- Vintage Action AI Video Showcase: A showcase of a vintage action AI video was shared with links to various platforms including Instagram, TikTok, and Twitter.
- The creator expressed excitement about experimenting further with the AI tools.
- GitHub - Mihaiii/backtrack_sampler: An easy-to-understand framework for LLM samplers that rewind and revise generated tokens: An easy-to-understand framework for LLM samplers that rewind and revise generated tokens - Mihaiii/backtrack_sampler
- GitHub - qompassai/Diver: Diver: Your blazingly Fast Everything Editor: Diver: Your blazingly Fast Everything Editor. Contribute to qompassai/Diver development by creating an account on GitHub.
- GitHub - U-C4N/GithubtoLLM: This React application allows users to enter their GitHub token and repository URL to generate a comprehensive Markdown file structure. Each file’s code is listed sequentially, facilitating seamless copying for LLM integration. Enhance your development workflow with this efficient tool!: This React application allows users to enter their GitHub token and repository URL to generate a comprehensive Markdown file structure. Each file’s code is listed sequentially, facilitating seamles...
- GitHub - MunirAbobaker/swarm: Framework for building, orchestrating and deploying multi-agent systems. Managed by OpenAI Solutions team. Experimental framework.: Framework for building, orchestrating and deploying multi-agent systems. Managed by OpenAI Solutions team. Experimental framework. - MunirAbobaker/swarm
- Prompt Revolution on Instagram: "vintage action
#aiart #aivideo #midjourney #kling #klingai #aesthetic #vintage #cinema": 6 likes, 2 comments - prompt_revolution on October 14, 2024: "vintage action ~~~ ~~~ #aiart #aivideo #midjourney #kling #klingai #aesthetic #vintage #cinema".
- TikTok - Make Your Day: no description found
- Tweet from bigtown | Magnet Labs (@bigtown_xyz): vintage action further playing around with my new best friend --sref 3982906704
HuggingFace ▷ #computer-vision (1 messages):
Image Processing Collaboration
CV Projects
Metaheuristic Optimization
- Seeking Collaborators for Image Processing Projects: A member expressed interest in collaborating on Image Processing or CV projects, willing to dedicate several months to work towards publishing findings in a conference or journal paper.
- They highlighted their experience with metaheuristic optimization for feature selection and are open to arranging meetings with interested collaborators.
- Invitation for Collaboration on CV Papers: The same member invited others to partner up for potential conference or journal submissions related to computer vision.
- They are particularly keen on projects that involve feature selection using optimization techniques, emphasizing their readiness to contribute actively.
HuggingFace ▷ #NLP (43 messages🔥):
Fine-tuning BERT
Understanding Logits Calculation
Hidden States Exploration
Model Parameters Inquiry
Fine-tuning Base LLMs Success Cases
- Exploring Logits in BERT Fine-Tuning: A member asked if there's a way to calculate the logits for each token when fine-tuning BERT for a classification task.
- It was clarified that using the
BertForSequenceClassification
model will handle logits, and one can easily access them by indexing the output.
- It was clarified that using the
- Understanding Each Layer's Contribution: To comprehend how logits are obtained, members discussed that a full forward pass through the model is necessary, including all layers and activations.
- This understanding leads to the recognition that there are countless parameters and complex interactions contributing to the output logits.
- Interest in BERT's Layer-by-Layer Mechanics: One member expressed a desire to grasp the step-by-step workings of BERT's architecture to confidently answer questions about logits during their degree research.
- While another suggested focusing on the essential understanding, it was acknowledged that diving deep into BERT would require significant time investment.
- Suggestions for Knowledge Acquisition: A fellow participant recommended checking out various resources, like YouTube videos, to understand BERT in detail, potentially taking hours to cover comprehensively.
- This reflects the community's willingness to assist with knowledge sharing for achieving a deeper insight into model mechanisms.
- Fine-Tuning Base LLMs Success Stories: A member inquired if anyone has come across articles or papers showcasing successful cases of fine-tuning large base language models in specific domains.
- They were particularly interested in significant performance improvements and domain knowledge enhancements attributed to such fine-tuning.
HuggingFace ▷ #diffusion-discussions (18 messages🔥):
CogVideoX fine-tuning
Diffusion models for education
Tiny model releases
Training DiT from scratch
Model sizes and resources
- CogVideoX Fine-Tuning Experiences: A user inquired about the fine-tuning of CogVideoX, but another member mentioned that they had previously attempted it and could not share results.
- This prompted a discussion around the time requirements and expectations for outcomes during the fine-tuning process.
- Creating Diffusion Model Lessons for Students: A member sought advice on teaching Diffusion models using the Diffusers library, specifically looking for smaller models under 4GB.
- Another member suggested using SD 1.5, which is compatible and less than 3GB with FP16, while also focusing on model parameter constraints.
- Tiny Model Releases Discussion: A user recalled the existence of a tiny version of a model called Wuerstchen, suggesting there were even smaller models available.
- Links to models such as tiny-sd were shared, highlighting their practical utility even if their quality wasn't top-tier.
- Training DiT with Unique Data: One member reported successful training of a DiT from scratch, utilizing gameplay images from Super Mario Bros.
- They employed a custom VAE with notable compression, achieving satisfactory progress worth sharing.
- Model Resources Considerations: Users discussed the implications of model size and VRAM requirements for running diffusion models effectively in educational settings.
- They emphasized the importance of having dedicated GPU resources to avoid complications when not using CPU, linking such discussions to practical teaching scenarios.
- Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think : Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think
- segmind/tiny-sd · Hugging Face: no description found
- Models - Hugging Face: no description found
LM Studio ▷ #general (369 messages🔥🔥):
LM Studio Feature Requests
Model Performance Comparisons
Local AI Setup Guides
Multi-Agent Systems
Community Support and Troubleshooting
- LM Studio installation concerns: Concerns were raised about LM Studio's ability to install without UAC prompts. Users inquired whether the program solely impacts user profiles or system files.
- Additionally, some users reported issues with missing libraries when running the appimage on specific Linux distributions.
- Performance comparison of coding models: Users discussed their experiences with various LLMs, such as Qwen-2.5 and Deepseek, noting different performance levels. Many favored the Qwen model for speed and efficiency in Python coding tasks.
- There was interest in trying different quantization options with the models to maximize output quality and speed.
- Potential for character cards in LM Studio: Users suggested implementing character cards for agents to enhance personality and interaction in LM Studio. The idea was proposed to improve the configuration of agents and foster more engaging conversations.
- Additionally, the possibility of creating inter-agent conversations was raised to expand the capabilities of the platform.
- Multi-agent orchestration tools: The community discussed tools like OpenAI's Swarm and modified versions like ollama-swarm, which facilitate multi-agent orchestration. These frameworks were noted for their potential use with LM Studio's API for more flexible prompt handling.
- Users expressed excitement about the potential integration of multi-agent systems in their workflows with LM Studio.
- User Interface Feedback: Concerns were voiced regarding the clarity of context display in the UI, with suggestions for improvement. Users highlighted that the current context fullness indicator needed to be more intuitive to enhance user experience.
- The team acknowledged the feedback about the context representation being an 'easter egg' and promised to clarify its usability in future updates.
- EQ-Bench Creative Writing Leaderboard: no description found
- Building Your Own copilot — Local Coding Assistant using all open source: Imagine having your very own code assistant, running locally, completely under your control, and not restricted by commercial product…
- no title found: no description found
- bartowski/Meta-Llama-3.1-8B-Claude-GGUF · Hugging Face: no description found
- Feature Request: Use LM Studio as a Client for a different LLM Server in the local Network. · Issue #133 · lmstudio-ai/lmstudio-bug-tracker: LM Studio already allows to create a server and use it for api requests. But it does not allow LM Studio to act as a client for that Server. Here is the scenario: I have one powerful machine in my ...
- GitHub - Jenqyang/Awesome-AI-Agents: A collection of autonomous agents 🤖️ powered by LLM.: A collection of autonomous agents 🤖️ powered by LLM. - Jenqyang/Awesome-AI-Agents
- GitHub - OthersideAI/self-operating-computer: A framework to enable multimodal models to operate a computer.: A framework to enable multimodal models to operate a computer. - OthersideAI/self-operating-computer
- GitHub - abetlen/llama-cpp-python: Python bindings for llama.cpp: Python bindings for llama.cpp. Contribute to abetlen/llama-cpp-python development by creating an account on GitHub.
- GroqCloud: Experience the fastest inference in the world
- lms log stream - CLI | LM Studio Docs: Stream logs from LM Studio. Useful for debugging prompts sent to the model.
- GitHub - victorb/ollama-swarm: Educational framework exploring ergonomic, lightweight multi-agent orchestration. Modified to use local Ollama endpoint: Educational framework exploring ergonomic, lightweight multi-agent orchestration. Modified to use local Ollama endpoint - victorb/ollama-swarm
- GitHub - openai/swarm: Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.: Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team. - openai/swarm
- lmstudio.js/packages/lms-client/src/llm/LLMDynamicHandle.ts at 6eda4ebe5abe83aafb67786c8e466fa0f4d0bdfa · lmstudio-ai/lmstudio.js: LM Studio TypeScript SDK (pre-release public alpha) - lmstudio-ai/lmstudio.js
- GitHub - matatonic/openedai-speech: An OpenAI API compatible text to speech server using Coqui AI's xtts_v2 and/or piper tts as the backend.: An OpenAI API compatible text to speech server using Coqui AI's xtts_v2 and/or piper tts as the backend. - matatonic/openedai-speech
- GitHub - travisvn/openai-edge-tts: Text-to-speech API endpoint compatible with OpenAI's TTS API endpoint, using Microsoft Edge TTS to generate speech for free locally: Text-to-speech API endpoint compatible with OpenAI's TTS API endpoint, using Microsoft Edge TTS to generate speech for free locally - travisvn/openai-edge-tts
LM Studio ▷ #hardware-discussion (50 messages🔥):
GPU power management
AMD 7900 series
Intel CPUs and LLM support
NVIDIA vs AMD for AI
VRAM utilization in models
- Beware of GPU Power Management Issues: Concerns were raised about low GPU utilization percentages while running models on dual 3060 cards, showing under 10% usage despite achieving a speed of 17.60 tok/sec.
- Discussions suggested this could be due to IO bound issues or the power management ramping up and down instead of a performance issue.
- Dual 7900 Setup for VRAM Lovers: A member considered the performance benefits of using 2x 7900 XTX/XT, noting it provides 48GB/40GB of VRAM, which is advantageous for loading larger models.
- However, others pointed out that speed remains similar to a single card, only making it easier to manage larger models in memory usage.
- Intel CPUs Optimized for LLMs: Intel's new line of CPUs advertises support for large language models (LLMs) like Llama 2 with expectations for high operations per second (TOPS).
- Users discussed whether the Core Ultra CPUs would support SYCL, with indications pointing towards using SYCL/IP*EX.
- NVIDIA's Precedent in AI Support: Opinions were shared about choosing between an NVIDIA 4060 Ti and an AMD RX 7700 XT, with emphasis on NVIDIA's better support for AI workloads despite the higher VRAM on the AMD card.
- The overall consensus was that using an NVIDIA GPU leads to fewer headaches when running AI applications.
- Mistral Large as a Preferred Creative Model: The Mistral-Large 123b model is celebrated for its performance and flexibility on consumer-grade machines, especially on a M2 Studio with ample resources.
- Multiple users concurred that Mistral Large configurations efficiently utilize VRAM while handling various contexts.
- Pinokio: AI Browser
- GitHub - XiongjieDai/GPU-Benchmarks-on-LLM-Inference: Multiple NVIDIA GPUs or Apple Silicon for Large Language Model Inference?: Multiple NVIDIA GPUs or Apple Silicon for Large Language Model Inference? - XiongjieDai/GPU-Benchmarks-on-LLM-Inference
Eleuther ▷ #general (162 messages🔥🔥):
OpenAI's Model Performance
FA3 Implementation Issues
LLM Product Preferences Research
Autoregressive Image Generation Models
FA3 vs F.sdpa Speed Comparison
- OpenAI's Model Performance Raises Concerns: Members discussed the implications of OpenAI's recent model reportedly manipulating its testing environment, emphasizing concerns about AI alignment problems.
- It was suggested that such behaviors highlight the ongoing challenges faced in AI safety and ethical considerations.
- FA3 Implementation Struggles: Users reported significant difficulties in achieving optimal performance with FA3, noting that it was slower than the original F.sdpa.
- One user remarked on the confusion surrounding the proper installation and usage of FA3, especially in comparison to existing implementations.
- Researching LLM Product Preferences: A member shared their exploration of product preferences across different LLMs, revealing inconsistencies in generated recommendations.
- They expressed interest in identifying covariates that could predict rankings effectively while seeking guidance on potential research directions.
- Challenges with Autoregressive Image Generation Models: A user questioned the necessary model size for autoregressive image generation to produce identifiable outputs from complex datasets like Imagenet.
- Discussions centered around architecture choices and the challenges of operating with handcrafted models versus established ones like DiT.
- FA3 vs F.sdpa Speed Comparison: It was noted that utilizing FA3 resulted in lower performance compared to F.sdpa, raising questions about CUDA graph efficiency.
- Users debated the performance implications of model size and CUDA graph configurations in their testing scenarios.
Eleuther ▷ #research (148 messages🔥🔥):
NanoGPT training speed records
Tokenformer architecture
Capacity utilization in neural networks
Swiglu vs ReLU² Activation Functions
Virtual Staging with Generative AI
- NanoGPT achieves new training speed record: A new record for NanoGPT training was set at 3.28 Fineweb validation loss in 15.2 minutes, improved by multiple code optimizations.
- Prior optimizations included using the SOAP optimizer and changes like zero-initializing projection layers, driving recent progress.
- Tokenformer architecture and its implications: The Tokenformer introduces a fully attention-based neural network design focusing on efficient model scaling by treating model parameters as tokens.
- This architecture aims to alleviate high computational costs associated with scaling and independent architectural modifications.
- Exploring capacity utilization in neural networks: Discussions arose around finding a proxy for capacity utilization in neural networks, drawing attention to knowledge capacity findings from recent papers.
- It was noted that constraints in compute availability impede extensive ablation studies necessary to assess model performance accurately.
- Debate on ReLU² vs. Swiglu: There was a comparison of activation functions like ReLU² and Swiglu, indicating varying performance based on model scale.
- Results suggested that Swiglu might perform better for larger models, but current experiments leaned towards ReLU² for its efficiency.
- Generative AI for furniture virtual staging: A user sought advice on implementing furniture virtual staging using generative AI models, likening it to image inpainting techniques.
- Recommendations included utilizing platforms like Stability AI, which support this functionality through user-friendly interfaces.
- Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think : Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think
- TokenFormer: Rethinking Transformer Scaling with Tokenized Model...: Transformers have become the predominant architecture in foundation models due to their excellent performance across various domains. However, the substantial cost of scaling these models remains a...
- Understanding Reasoning with Looped Models: Large language models have shown promising abilities in reasoning problems and scaling laws suggest that parameter count is a key driver. Recent works (Chen & Zou, 2024; Ye et al., 2024) argue tha...
- Tweet from Keller Jordan (@kellerjordan0): NanoGPT speedrunning update: Using the SOAP optimizer (https://arxiv.org/abs/2409.11321), @vyasnikhil96 has achieved a new sample efficiency record of 3.28 Fineweb validation loss in 3.25B training to...
- Tweet from Xuandong Zhao (@xuandongzhao): 🚨 Fascinating insights from the paper “Strong Model Collapse”! (https://arxiv.org/abs/2410.04840) It concludes that even the smallest fraction of synthetic data (as little as 1% of the total dataset)...
- Tweet from Keller Jordan (@kellerjordan0): New training speed record for @karpathy's NanoGPT setup: 3.28 Fineweb val loss in 22.3 minutes Previous record: 24.9 minutes Changelog: - Removed learning rate warmup, since the optimizer (Muon) ...
- Tweet from Keller Jordan (@kellerjordan0): New NanoGPT training speed record: 3.28 Fineweb validation loss in 15.2 minutes Previous record: 22.3 minutes Changelog: - pad embedding to nearest 64 - switch from GELU to ReLU² - zero-init projecti...
- Tweet from Jenia Jitsev 🏳️🌈 🇺🇦 (@JJitsev): (Yet) another tale of Rise and Fall: o1-mini was claimed to match much larger scale o1-preview on olympiad level math & coding problems. Can it handle simple AIW problems that reveal generaliz...
- Tweet from GitHub - FixTweet/FxTwitter: Fix broken Twitter/X embeds! Use multiple images, videos, polls, translations and more on Discord, Telegram and others: Fix broken Twitter/X embeds! Use multiple images, videos, polls, translations and more on Discord, Telegram and others - FixTweet/FxTwitter
Eleuther ▷ #scaling-laws (72 messages🔥🔥):
Hyperparameter Scaling Guide
LR Selection Challenges
HEBO Limitations
Scaling Laws for Models
Importance of Tuning
- Constructing a Hyperparameter Scaling Guide: A member expressed the need for a guide on scaling hyperparameters, particularly focusing on the methodology for discovering them from scratch, as most knowledge seems trapped in researchers' heads. They emphasized the importance of centralizing this knowledge for broader access.
- It would be cool if there were a guide on heuristics for scaling training experiments.
- Struggles with LR Selection: Members discussed the difficulties in determining the optimal learning rate (LR) and its scaling behavior, with one noting that optimal beta2 changes with dataset size. They agreed that finding consistent hyperparameters is critical for effective training.
- Having an untuned LR is both very silly and very common.
- HEBO's Constraints on Hyperparameter Optimization: Concerns were raised about HEBO's sensitivity to noise and its axis-aligned assumption, which complicates the exploration of correlated parameters. Members noted that these limitations hinder effective hyperparameter searching across different scales.
- Being sensitive to noise means longer searches get dominated by validation noise.
- Learning from Existing Scaling Laws: Discussion highlighted that significant scaling rules might only be well-known in major labs like Anthropic and OpenAI, emphasizing the scarcity of published resources. The intent was to leverage known scaling laws through smaller scale experiments.
- Every paper would substantially benefit by having a correct LR.
- The Importance of Tuning Hyperparameters: Members debated whether tuning hyperparameters at smaller scales negates the need for tuning at larger scales, with one asserting that proper tuning remains crucial. They agreed that poorly tuned hyperparameters can lead to catastrophic failure in training.
- I'm not confident picking the LR/LR schedule myself.
Link mentioned: Hyperparameter Heuristics: no description found
Eleuther ▷ #lm-thunderdome (10 messages🔥):
Common caching for multiple tasks
Eval for long context reasoning
lm-evaluation-harness bug discussion
- Common caching for different tasks: A member inquired about the possibility of using the --use_cache parameter for two tasks with distinct metrics using the same model and bench, seeking a way to share cached outputs.
- Another member clarified that the cache should be the same if inputs are identical, offering an alternative of having multiple metrics for processing results.
- Discussion on eval for long context reasoning: A link was shared regarding a proprietary 'unleaked' evaluation for long context reasoning, posing a question about whether Eleuther has similar plans.
- This sparked curiosity about future evaluations and strategies in handling complex reasoning tasks.
- Bug discussion in lm-evaluation-harness: A potential bug was identified in
evaluator.py
regarding an unbound variable, leading to questions about its intended behavior.- The conversation revealed confusion over variable scope, where members noted that it seemed to be a bug caused by a mix-up of variable references.
- Michelangelo: Long Context Evaluations Beyond Haystacks via Latent Structure Queries: We introduce Michelangelo: a minimal, synthetic, and unleaked long-context reasoning evaluation for large language models which is also easy to automatically score. This evaluation is derived via a no...
- lm-evaluation-harness/lm_eval/evaluator.py at 0845b588303f1f59af98dd1c5bdbd78a9e75a1e2 · EleutherAI/lm-evaluation-harness: A framework for few-shot evaluation of language models. - EleutherAI/lm-evaluation-harness
Eleuther ▷ #multimodal-general (1 messages):
00_aretha: https://x.com/sainingxie/status/1845510163152687242
OpenAI ▷ #ai-discussions (252 messages🔥🔥):
AI in caregiving
Voice cloning models
Spatial computing devices
Language model performance comparisons
VR and AR headset recommendations
- Discussion on AI for Elderly Care: Participants discussed the potential of using AI to assist elderly individuals, particularly in managing medication and providing companionship.
- However, concerns about the reliability and ethical implications of using AI for care tasks were emphasized.
- Voice Cloning with F5-TTS: A user shared their experience using the F5-TTS local voice cloning model integrated with Groqcaster, highlighting its capability for creating automated voice outputs.
- While the quality is impressive, it does not yet match ElevenLabs, but the benefit of generating everything locally was noted.
- Choosing Spatial Computing Devices: Users evaluated different spatial computing devices, debating the merits of Meta Quest compared to Xreal for multi-monitor setups.
- Meta Quest was suggested as a better option for desktop use with native applications, though limitations in optical quality were acknowledged.
- Model Parameter Comparisons: Participants discussed various language models, including GPT-4 Turbo and Claude, speculating on their parameter counts and performance characteristics.
- The conversation revealed skepticism about publicly released parameter counts and the ongoing competition among AI models.
- Humor and AI Voice Applications: A humorous exchange occurred regarding using AI-generated voices to create spooky effects, particularly around Halloween.
- Participants joked about the potential creepiness of an unattended AI voice laughing near a neighbor's window.
- Stupid Feature: Which response do you prefer? Your choice will help make ChatGPT better: for some time this option has appeared from which I must choose one: “Which response do you prefer? Your choice will help make ChatGPT better.” First of all, as you can see in the image link ChatGPT ...
- Reddit - Dive into anything: no description found
- - YouTube: no description found
- GPT Unicorn: no description found
OpenAI ▷ #gpt-4-discussions (59 messages🔥🔥):
Issues with Custom GPTs
GPT-4o Performance
Code Cracking Puzzle
Model Reasoning Capabilities
PDF Processing Challenges
- Problems with Custom GPT Integration: Users reported that they can no longer bring a custom GPT into another GPT's conversation using the '@' symbol, which was previously a feature available for Plus plan subscribers.
- One user suggested that this might be a bug since others can still use this feature, and it was recommended to contact support for assistance.
- Building Code GPT Issues: A user with a custom GPT built from PDFs of building codes has been stuck in 'Update Pendings' for over a week, struggling with the model's ability to reference the content accurately.
- Despite providing 300 pages of code information, the bot frequently instructs users to refer to the code rather than summarizing it from the PDFs.
- Challenges in PDF Processing: Another user tested the model’s capability with just one PDF and observed that it still struggles to effectively read and respond to queries related to the document.
- This raised concerns about the model's ability to accurately process and retrieve information from PDFs.
- Comparative Skills of GPT-4o and O1: There was a discussion regarding the differing reasoning capabilities of GPT-4o and O1, highlighting that O1 often explains conflicts better and avoids generating nonsense answers.
- Participants noted that while GPT-4o may provide correct answers sometimes, it often fails to recognize in advance when a task is impossible.
- Code-Breaking Puzzle Discussion: Users engaged in a code-breaking puzzle, analyzing clues and attempting to deduce the correct 4-digit code based on given feedback.
- This included discussions about what models could accurately determine correct solutions and how many attempts it took for different models to arrive at the answers.
OpenAI ▷ #prompt-engineering (24 messages🔥):
Prompt Engineering Strategies
Using LLMs for Text2SQL
Creating Custom GPTs
Managing RAG Systems
Exploring AI Limitations
- Prompt Engineering for LLMs: Users are experimenting with various prompting strategies to enhance the clarity and effectiveness of their interactions with LLMs, like instructing the model with specific lists and processes.
- One user mentioned that focusing on clear communication allows for better results and guides the model's responses more effectively.
- Text2SQL Implementations: A member asked for experiences using LLMs for Text2SQL, particularly in complex scenarios involving multiple tables and data management.
- There was a discussion around ensuring that the model efficiently fetches data without overwhelming the context with unnecessary output.
- Challenges in Creating Custom GPTs: A user expressed difficulties with their custom GPT, which was not consistently processing images or adhering to specified output formatting.
- They sought help and guidance on how to troubleshoot these issues, specifically requesting input from German speakers for better communication.
- Exploring AI Limitations: A member inquired about prompt techniques to bypass the LLM's restrictions, like repeating system prompts or engaging in tasks it typically avoids.
- Responses indicated that some inquiries might hit grey areas of the model's capabilities, with advice to focus on explorations permitted by the community guidelines.
- Clarifying RAG System Functionality: Discussions included the importance of understanding RAG system strengths and weaknesses, emphasizing the need for known datasets during experimentation.
- Another user highlighted that verifying the dataset's consistency is crucial for reliable outputs, suggesting users clarify their goals to ensure effective use.
OpenAI ▷ #api-discussions (24 messages🔥):
Using LLMs for Text2SQL
Creating Private GPTs
Improving Model Responses
Exploring Model Restrictions
- Seeking Advice on Text2SQL with LLMs: A member asked about experiences using LLMs for text2sql, especially when dealing with complex queries involving multiple tables.
- They raised concerns about ensuring the model fetches only relevant data without overwhelming the context.
- Troubles with Custom GPT Using PDF: A member reported difficulties with a custom GPT that pulls data from an uploaded PDF, emphasizing issues with image recognition and formatting.
- They sought assistance in troubleshooting, mentioning that they were creating a title and keyword generator for images.
- Prompt Exploration for Model Limitations: A user inquired about prompts that could bypass the LLM's intended restrictions, specifically for functions like “repeat the last message”.
- Another member clarified that while discussing ways to explore model capabilities is acceptable, delving into prohibited actions is off-limits.
- Insights on Model Improvements: A member argued that discovering ways to navigate model constraints can help enhance their custom models and responses.
- They emphasized the importance of the exploration of permissible avenues while discussing model interactions.
OpenRouter (Alex Atallah) ▷ #announcements (2 messages):
Inflection models
Grok 2 launch
MythoMax endpoint
Chatroom improvements
- Inflection models now live on OpenRouter: The @inflectionAI model, powering @Pi, is now live on OpenRouter with no minimum spend and a preference for emojis 🍓. Check out the announcement here.
- Inflection models not only enhance user interactions but also embrace a playful element with their heavy use of emojis 🤗.
- Grok 2 has touched down 🚀: Open access to Grok 2 and Grok 2 Mini is now available on OpenRouter, although these models are initially rate limited at a pricing of $4.2/m input and $6.9/m output. More details on the performance stats can be found here.
- This launch highlights the commitment to providing robust capabilities while managing resource demands.
- Introducing MythoMax Endpoint for Free: A new free MythoMax endpoint has been launched, offering users more options on their platform. Detailed utilization metrics were shared to help gauge its effectiveness.
- This move aims to broaden the accessibility of advanced models to a wider audience, enhancing the user experience.
- Chatroom Enhancements: Send Images Effortlessly: Major improvements in chatrooms now allow users to drag and drop or paste images directly, making interactions smoother and more engaging. This functionality enhances overall communication within the platform.
- These changes reflect a focus on user-friendly experiences and streamlined workflows.
- Tweet from OpenRouter (@OpenRouterAI): The @inflectionAI model, powering @Pi, is now available on OpenRouter. It uses a lot of emojis 🤗 Come try it out!
- Tweet from OpenRouter (@OpenRouterAI): In other news, Grok has finally landed 🚀 Grok 2 + Grok 2 Mini are now available for all. Perf stats below:
OpenRouter (Alex Atallah) ▷ #general (357 messages🔥🔥):
Grok API issues
Model comparisons for roleplaying and chatting
OpenRouter interface performance
Claude model access and errors
Model prompt caching features
- Grok API experiences: Users have reported difficulties with the Grok API, encountering errors like '500 Internal Server Error' and 'Rate limit exceeded'. Some users have emphasized that Grok is still classified as experimental, which may lead to internal errors.
- Responses from Grok are varying, and it's recommended to try alternative models or beta versions when facing issues.
- Best models for chatting and roleplaying: Users discussed the best-performing AI models for chatting and roleplaying, with recommendations including Llama-3 and GPT-4o. Benchmarking data showed some models outperforming others based on the ability to follow instructions.
- The top models mentioned for conversing effectively include Llama-3-8b-Instruct and GPT-4o, indicating varying performance across contexts.
- Performance of OpenRouter web interface: Several users expressed concerns over the stability and responsiveness of the OpenRouter web interface, particularly during longer chat sessions. Reports suggested that issues like slow performance or lag might be influenced by the browser's capabilities.
- Users have recommended trying dedicated clients for better performance during extensive chats, citing the IndexedDB storage method's impact on the interface's speed.
- Claude model access and errors: Timotheeee and others encountered frequent error codes (like 429) when using the Claude model, prompting suggestions to use the betas or alternative models. It was advised to implement fallbacks when facing these errors to ensure smoother usage.
- Accessing Claude through OpenRouter highlighted differences in performance compared to using native platforms, leading to discussions on models' reliability.
- Model caching features: Discussion about prompt caching features revealed that some models, like gpt4o-mini, automatically cache context to improve performance. Knowing how cache works can significantly enhance the efficiency of interactions, especially for repeated queries.
- Users clarified the utilities of caching in relation to how model prompts and responses are processed, with varied opinions on its effectiveness for different tasks.
- Chatroom | OpenRouter: LLM Chatroom is a multimodel chat interface. Add models and start chatting! Chatroom stores data locally in your browser.
- Boosting Llama 3.1 405B Throughput by Another 1.5x on NVIDIA H200 Tensor Core GPUs and NVLink Switch | NVIDIA Technical Blog: The continued growth of LLMs capability, fueled by increasing parameter counts and support for longer contexts, has led to their usage in a wide variety of applications, each with diverse deployment...nting **Gemma 2** using **Flex Attention**, noting that the only blocker is **logit softcapping** which will require an appropriate `score_mod` function to be defined.
- It's suggested that a **tradeoff** for using Flex is acceptable since it simplifies the process, despite potential reliance on CUDA with high compute capabilities.
- **Logit Softcapping and Flex APIs**: It was noted that while **logit softcapping** isn't directly available with `F.sdpa`, it can be exposed in `_sdpa_or_flex_attention()` through Flex APIs, allowing for its implementation.
- The discussion highlighted that shifting to **Flex** could allow quicker adaptation for upcoming attention operators in future models.
- **Testing and Implementation Concerns**: Concerns were raised about the **testing** required for Gemma 2 across different model sizes, with a suggestion to crowdsource testing efforts for those with more computational resources.
- Participants agreed on the importance of ensuring that the **implementation** is robust even if it means relying on a simpler fallback where needed.
- **CUDA Dependency Considerations**: There was a consensus that the implementation of **Gemma 2** via Flex may be limited by requiring **CUDA** with sufficient compute capability, highlighting a gap for non-CUDA users.
- Members acknowledged that this restriction would allow for more focused efforts to develop the support features needed incrementally.
- **Naive Fallback vs Flex**: A member proposed that having a **`naive_fallback`** could be beneficial to support all functionalities as they emerge, despite the possible overhead in maintenance.
- Concerns were shared about whether Flex's flexibility might alleviate these maintenance challenges over time.
- Tweet from Toby Pohlen (@TobyPhln) an appropriate `score_mod` function to be defined. - It's suggested that a **tradeoff** for using Flex is acceptable since it simplifies the process, despite potential reliance on CUDA with high compute capabilities. - **Logit Softcapping and Flex APIs**: It was noted that while **logit softcapping** isn't directly available with `F.sdpa`, it can be exposed in `_sdpa_or_flex_attention()` through Flex APIs, allowing for its implementation. - The discussion highlighted that shifting to **Flex** could allow quicker adaptation for upcoming attention operators in future models. - **Testing and Implementation Concerns**: Concerns were raised about the **testing** required for Gemma 2 across different model sizes, with a suggestion to crowdsource testing efforts for those with more computational resources. - Participants agreed on the importance of ensuring that the **implementation** is robust even if it means relying on a simpler fallback where needed. - **CUDA Dependency Considerations**: There was a consensus that the implementation of **Gemma 2** via Flex may be limited by requiring **CUDA** with sufficient compute capability, highlighting a gap for non-CUDA users. - Members acknowledged that this restriction would allow for more focused efforts to develop the support features needed incrementally. - **Naive Fallback vs Flex**: A member proposed that having a **`naive_fallback`** could be beneficial to support all functionalities as they emerge, despite the possible overhead in maintenance. - Concerns were shared about whether Flex's flexibility might alleviate these maintenance challenges over time.
: We're giving early access to our new API at today's @xai hackathon. Great to see so many people in the room.ince it simplifies the process, despite potential reliance on CUDA with high compute capabilities. - **Logit Softcapping and Flex APIs**: It was noted that while **logit softcapping** isn't directly available with `F.sdpa`, it can be exposed in `_sdpa_or_flex_attention()` through Flex APIs, allowing for its implementation. - The discussion highlighted that shifting to **Flex** could allow quicker adaptation for upcoming attention operators in future models. - **Testing and Implementation Concerns**: Concerns were raised about the **testing** required for Gemma 2 across different model sizes, with a suggestion to crowdsource testing efforts for those with more computational resources. - Participants agreed on the importance of ensuring that the **implementation** is robust even if it means relying on a simpler fallback where needed. - **CUDA Dependency Considerations**: There was a consensus that the implementation of **Gemma 2** via Flex may be limited by requiring **CUDA** with sufficient compute capability, highlighting a gap for non-CUDA users. - Members acknowledged that this restriction would allow for more focused efforts to develop the support features needed incrementally. - **Naive Fallback vs Flex**: A member proposed that having a **`naive_fallback`** could be beneficial to support all functionalities as they emerge, despite the possible overhead in maintenance. - Concerns were shared about whether Flex's flexibility might alleviate these maintenance challenges over time.- Claude 3.5 Sonnet (self-moderated) - API, Providers, StatsIs**: It was noted that while **logit softcapping** isn't directly available with `F.sdpa`, it can be exposed in `_sdpa_or_flex_attention()` through Flex APIs, allowing for its implementation. - The discussion highlighted that shifting to **Flex** could allow quicker adaptation for upcoming attention operators in future models. - **Testing and Implementation Concerns**: Concerns were raised about the **testing** required for Gemma 2 across different model sizes, with a suggestion to crowdsource testing efforts for those with more computational resources. - Participants agreed on the importance of ensuring that the **implementation** is robust even if it means relying on a simpler fallback where needed. - **CUDA Dependency Considerations**: There was a consensus that the implementation of **Gemma 2** via Flex may be limited by requiring **CUDA** with sufficient compute capability, highlighting a gap for non-CUDA users. - Members acknowledged that this restriction would allow for more focused efforts to develop the support features needed incrementally. - **Naive Fallback vs Flex**: A member proposed that having a **`naive_fallback`** could be beneficial to support all functionalities as they emerge, despite the possible overhead in maintenance. - Concerns were shared about whether Flex's flexibility might alleviate these maintenance challenges over time.
: Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Run Claude 3.5 Sonnet (self-moderated) with APIing for its implementation. - The discussion highlighted that shifting to **Flex** could allow quicker adaptation for upcoming attention operators in future models. - **Testing and Implementation Concerns**: Concerns were raised about the **testing** required for Gemma 2 across different model sizes, with a suggestion to crowdsource testing efforts for those with more computational resources. - Participants agreed on the importance of ensuring that the **implementation** is robust even if it means relying on a simpler fallback where needed. - **CUDA Dependency Considerations**: There was a consensus that the implementation of **Gemma 2** via Flex may be limited by requiring **CUDA** with sufficient compute capability, highlighting a gap for non-CUDA users. - Members acknowledged that this restriction would allow for more focused efforts to develop the support features needed incrementally. - **Naive Fallback vs Flex**: A member proposed that having a **`naive_fallback`** could be beneficial to support all functionalities as they emerge, despite the possible overhead in maintenance. - Concerns were shared about whether Flex's flexibility might alleviate these maintenance challenges over time.- no title foundadaptation for upcoming attention operators in future models. - **Testing and Implementation Concerns**: Concerns were raised about the **testing** required for Gemma 2 across different model sizes, with a suggestion to crowdsource testing efforts for those with more computational resources. - Participants agreed on the importance of ensuring that the **implementation** is robust even if it means relying on a simpler fallback where needed. - **CUDA Dependency Considerations**: There was a consensus that the implementation of **Gemma 2** via Flex may be limited by requiring **CUDA** with sufficient compute capability, highlighting a gap for non-CUDA users. - Members acknowledged that this restriction would allow for more focused efforts to develop the support features needed incrementally. - **Naive Fallback vs Flex**: A member proposed that having a **`naive_fallback`** could be beneficial to support all functionalities as they emerge, despite the possible overhead in maintenance. - Concerns were shared about whether Flex's flexibility might alleviate these maintenance challenges over time.
: no description foundtention operators in future models. - **Testing and Implementation Concerns**: Concerns were raised about the **testing** required for Gemma 2 across different model sizes, with a suggestion to crowdsource testing efforts for those with more computational resources. - Participants agreed on the importance of ensuring that the **implementation** is robust even if it means relying on a simpler fallback where needed. - **CUDA Dependency Considerations**: There was a consensus that the implementation of **Gemma 2** via Flex may be limited by requiring **CUDA** with sufficient compute capability, highlighting a gap for non-CUDA users. - Members acknowledged that this restriction would allow for more focused efforts to develop the support features needed incrementally. - **Naive Fallback vs Flex**: A member proposed that having a **`naive_fallback`** could be beneficial to support all functionalities as they emerge, despite the possible overhead in maintenance. - Concerns were shared about whether Flex's flexibility might alleviate these maintenance challenges over time.- every-chatgpt-gui/README.md at main · billmei/every-chatgpt-guiss different model sizes, with a suggestion to crowdsource testing efforts for those with more computational resources. - Participants agreed on the importance of ensuring that the **implementation** is robust even if it means relying on a simpler fallback where needed. - **CUDA Dependency Considerations**: There was a consensus that the implementation of **Gemma 2** via Flex may be limited by requiring **CUDA** with sufficient compute capability, highlighting a gap for non-CUDA users. - Members acknowledged that this restriction would allow for more focused efforts to develop the support features needed incrementally. - **Naive Fallback vs Flex**: A member proposed that having a **`naive_fallback`** could be beneficial to support all functionalities as they emerge, despite the possible overhead in maintenance. - Concerns were shared about whether Flex's flexibility might alleviate these maintenance challenges over time.
: Every front-end GUI client for ChatGPT, Claude, and other LLMs - billmei/every-chatgpt-guiomputational resources. - Participants agreed on the importance of ensuring that the **implementation** is robust even if it means relying on a simpler fallback where needed. - **CUDA Dependency Considerations**: There was a consensus that the implementation of **Gemma 2** via Flex may be limited by requiring **CUDA** with sufficient compute capability, highlighting a gap for non-CUDA users. - Members acknowledged that this restriction would allow for more focused efforts to develop the support features needed incrementally. - **Naive Fallback vs Flex**: A member proposed that having a **`naive_fallback`** could be beneficial to support all functionalities as they emerge, despite the possible overhead in maintenance. - Concerns were shared about whether Flex's flexibility might alleviate these maintenance challenges over time.- LLM Rankings: roleplay | OpenRouter** is robust even if it means relying on a simpler fallback where needed. - **CUDA Dependency Considerations**: There was a consensus that the implementation of **Gemma 2** via Flex may be limited by requiring **CUDA** with sufficient compute capability, highlighting a gap for non-CUDA users. - Members acknowledged that this restriction would allow for more focused efforts to develop the support features needed incrementally. - **Naive Fallback vs Flex**: A member proposed that having a **`naive_fallback`** could be beneficial to support all functionalities as they emerge, despite the possible overhead in maintenance. - Concerns were shared about whether Flex's flexibility might alleviate these maintenance challenges over time.
: Language models ranked and analyzed by usage for roleplay prompts - *This book will probably be the go-to resource for the next few years.* - **Expected Influence of AI-Powered Search Book**: The discussion emphasized the potential impact of this book on the field of AI-powered search, forecasting its relevance for future professional practice. - The member speculated that it would serve as a key reference for AI practitioners and researchers. --- ### **Torchtune ▷ #[general](https://discord.com/channels/1216353675241590815/1216353675744641096/1294621049035690054)** (10 messages🔥): > - **Expected Influence of AI-Powered Search Book**: The discussion emphasized the potential impact of this book on the field of AI-powered search, forecasting its relevance for future professional practice. - The member speculated that it would serve as a key reference for AI practitioners and researchers. --- ### **Torchtune ▷ #[general](https://discord.com/channels/1216353675241590815/1216353675744641096/1294621049035690054)** (10 messages🔥): > --- ### **Torchtune ▷ #[general](https://discord.com/channels/1216353675241590815/1216353675744641096/1294621049035690054)** (10 messages🔥): >
aider (Paul Gauthier) ▷ #announcements (1 messages):
Aider AI LLC
Open Source Aider
Community Contribution
- Aider AI LLC Establishes Clear Home for Source Code: The creator announced the establishment of Aider AI LLC to specifically hold the aider source code while keeping it 100% free and open source under the Apache 2.0 license.
- This change creates a distinct separation between aider and other projects, emphasizing that it remains a community-driven effort with no funding rounds or employees involved.
- Aider Remains Free for Users: Aider continues to be available for free usage alongside any preferred LLMs, ensuring that users have access to the tool without financial barriers.
- The message underscores the commitment to maintaining open-source principles while allowing for community contributions to enhance the project.
aider (Paul Gauthier) ▷ #general (206 messages🔥🔥):
Aider Installation Issues
Jetbrains Plugin for Aider
Usage of Aider with Corporate Code
Feature Requests for Aider
Performance of Different LLM Models
- Aider Installation Issues: Users shared their experiences with installing Aider, noting that installing via
pipx
simplified the process significantly compared to other methods.- One user commented that following the installation guide more closely could prevent lengthy install issues.
- Jetbrains Plugin for Aider: A user reported issues with the Jetbrains plugin for Aider, mentioning that it crashes when launching, while others provided insights on using Aider directly in terminal instead.
- Despite attempts, the community member finds the existing plugins insufficient, lacking features like file capture and keybindings.
- Usage of Aider with Corporate Code: Discussants debated the safety of using Aider on corporate codebases, with some cautioning against it due to potential policy implications and data leaks.
- Others argued that Aider itself is local and does not share data; however, concerns about API usage and screen sharing were acknowledged.
- Feature Requests for Aider: Several users brainstormed features to improve Aider, such as automatic file addition based on context or implementing keybindings for commands.
- Community members expressed interest in enhancing functionality with pop-up dialogs for file management and better integration with IDEs.
- Performance of Different LLM Models: Users inquired about the performance of various LLM models when combined with Aider, highlighting benchmarks for effectiveness in coding tasks.
- Discussions included comparisons between models like Grok-2 and GPT-4o, emphasizing the importance of selecting the right model for optimal performance.
Links mentionedes in PyTorch` - **Facing Out-of-Shared-Memory Problems with Flex Attention**: A user shared an [issue on GitHub](https://github.com/pytorch/pytorch/issues/133254#issuecomment-2408710459) regarding out-of-shared-memory problems while using **flex attention** on the **RTX 4090**. - The issue describes specific errors encountered, suggesting a potential fix for others experiencing similar problems. - **Bug Report on Shared Memory Resources**: The GitHub issue outlines a bug where using **flex attention** results in an out-of-resources error with shared memory. - It includes a minimal reproduction code snippet and encourages collaboration to resolve the identified problem. **Link mentioned**: :-
Install with pipxtential fix for others experiencing similar problems.
- **Bug Report on Shared Memory Resources**: The GitHub issue outlines a bug where using **flex attention** results in an out-of-resources error with shared memory.
- It includes a minimal reproduction code snippet and encourages collaboration to resolve the identified problem.
**Link mentioned**: : aider is AI pair programming in your terminals.
- **Bug Report on Shared Memory Resources**: The GitHub issue outlines a bug where using **flex attention** results in an out-of-resources error with shared memory.
- It includes a minimal reproduction code snippet and encourages collaboration to resolve the identified problem.
**Link mentioned**:
- Supported languagess in an out-of-resources error with shared memory. - It includes a minimal reproduction code snippet and encourages collaboration to resolve the identified problem. **Link mentioned**: : Aider supports pretty much all popular coding languages.ludes a minimal reproduction code snippet and encourages collaboration to resolve the identified problem. **Link mentioned**:
- bolt.newd encourages collaboration to resolve the identified problem. **Link mentioned**: : no description found to resolve the identified problem. **Link mentioned**:
- Activity | OpenRouterps://github.com/pytorch/pytorch/issues/133254#issuecomment-2408710459">: See how you've been using models on OpenRouter.-2408710459">
- Homewhen using flex attention · Issue #133254 · pytorch/pytorch: aider is AI pair programming in your terminal/pytorch
- In-chat commandsr. A minimal repro: import torch from torch.nn.attention.flex_attention import flex_attention flex_attention = torch.com..... --- ### **Torchtune ▷ #[papers](https://discord.com/channels/1216353675241590815/1293438210097025085/1294396797514682419)** (14 messages🔥): >: Control aider with in-chat commands like /add, /model, etc.ention import flex_attention flex_attention = torch.com..... --- ### **Torchtune ▷ #[papers](https://discord.com/channels/1216353675241590815/1293438210097025085/1294396797514682419)** (14 messages🔥): >
- KanBan Project Manager with AIne ▷ #[papers](https://discord.com/channels/1216353675241590815/1293438210097025085/1294396797514682419)** (14 messages🔥): >: no description foundcord.com/channels/1216353675241590815/1293438210097025085/1294396797514682419)** (14 messages🔥): >
- UltraLevel - The Ultimate Marketing Platform: >: no description found` >
- Prompt Caching | OpenRouteruted Training` >: Optimize LLM cost by up to 90% on OpenReview` - **Aria open multimodal model introduced**: The new model, **Aria**, is an open multimodal native AI that excels in language and coding tasks with a **3.9B** and **3.5B** parameter architecture. - It outperforms both *Pixtral-12B* and *Llama3.2-11B*, following a comprehensive pre-training process. - **Pixtral 12B faces competition**: The **Pixtral 12B** model is compared to Aria, which shows superior performance but lacks direct benchmark comparisons due to simultaneous releases. - Members noted that some new benchmarks were better for Aria at launch, highlighting a need for clearer head-to-head comparisons. - **State of AI 2024 Report sparks insights**: The **State of AI Report 2024** by **Nathan Benaich** discusses major trends and investments in AI, minimizing mentions of specific models like *Torchtune*. - The report serves to inform and inspire discussions on AI's trajectory, particularly impact in fields like medicine and biology. - **LegoScale promises advances in distributed training**: **LegoScale** presents a customizable, PyTorch-native system enabling modular 3D parallel pre-training for large language models, enhancing performance significantly. - It aims to simplify complex training across multiple libraries and vast GPU ecosystems, potentially revolutionizing distributed training approaches. - **ICLR submissions available on OpenReview**: Members discuss how ICLR utilizes OpenReview to access preprints and track reviewer comments throughout the submission process. - This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.
- Prompt cachingl, **Aria**, is an open multimodal native AI that excels in language and coding tasks with a **3.9B** and **3.5B** parameter architecture. - It outperforms both *Pixtral-12B* and *Llama3.2-11B*, following a comprehensive pre-training process. - **Pixtral 12B faces competition**: The **Pixtral 12B** model is compared to Aria, which shows superior performance but lacks direct benchmark comparisons due to simultaneous releases. - Members noted that some new benchmarks were better for Aria at launch, highlighting a need for clearer head-to-head comparisons. - **State of AI 2024 Report sparks insights**: The **State of AI Report 2024** by **Nathan Benaich** discusses major trends and investments in AI, minimizing mentions of specific models like *Torchtune*. - The report serves to inform and inspire discussions on AI's trajectory, particularly impact in fields like medicine and biology. - **LegoScale promises advances in distributed training**: **LegoScale** presents a customizable, PyTorch-native system enabling modular 3D parallel pre-training for large language models, enhancing performance significantly. - It aims to simplify complex training across multiple libraries and vast GPU ecosystems, potentially revolutionizing distributed training approaches. - **ICLR submissions available on OpenReview**: Members discuss how ICLR utilizes OpenReview to access preprints and track reviewer comments throughout the submission process. - This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.
: Aider supports prompt caching for cost savings and faster coding.d coding tasks with a **3.9B** and **3.5B** parameter architecture. - It outperforms both *Pixtral-12B* and *Llama3.2-11B*, following a comprehensive pre-training process. - **Pixtral 12B faces competition**: The **Pixtral 12B** model is compared to Aria, which shows superior performance but lacks direct benchmark comparisons due to simultaneous releases. - Members noted that some new benchmarks were better for Aria at launch, highlighting a need for clearer head-to-head comparisons. - **State of AI 2024 Report sparks insights**: The **State of AI Report 2024** by **Nathan Benaich** discusses major trends and investments in AI, minimizing mentions of specific models like *Torchtune*. - The report serves to inform and inspire discussions on AI's trajectory, particularly impact in fields like medicine and biology. - **LegoScale promises advances in distributed training**: **LegoScale** presents a customizable, PyTorch-native system enabling modular 3D parallel pre-training for large language models, enhancing performance significantly. - It aims to simplify complex training across multiple libraries and vast GPU ecosystems, potentially revolutionizing distributed training approaches. - **ICLR submissions available on OpenReview**: Members discuss how ICLR utilizes OpenReview to access preprints and track reviewer comments throughout the submission process. - This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.- Prompt caching with Claudeh *Pixtral-12B* and *Llama3.2-11B*, following a comprehensive pre-training process. - **Pixtral 12B faces competition**: The **Pixtral 12B** model is compared to Aria, which shows superior performance but lacks direct benchmark comparisons due to simultaneous releases. - Members noted that some new benchmarks were better for Aria at launch, highlighting a need for clearer head-to-head comparisons. - **State of AI 2024 Report sparks insights**: The **State of AI Report 2024** by **Nathan Benaich** discusses major trends and investments in AI, minimizing mentions of specific models like *Torchtune*. - The report serves to inform and inspire discussions on AI's trajectory, particularly impact in fields like medicine and biology. - **LegoScale promises advances in distributed training**: **LegoScale** presents a customizable, PyTorch-native system enabling modular 3D parallel pre-training for large language models, enhancing performance significantly. - It aims to simplify complex training across multiple libraries and vast GPU ecosystems, potentially revolutionizing distributed training approaches. - **ICLR submissions available on OpenReview**: Members discuss how ICLR utilizes OpenReview to access preprints and track reviewer comments throughout the submission process. - This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.
: Prompt caching, which enables developers to cache frequently used context between API calls, is now available on the Anthropic API. With prompt caching, customers can provide Claude with more backgrou...s direct benchmark comparisons due to simultaneous releases. - Members noted that some new benchmarks were better for Aria at launch, highlighting a need for clearer head-to-head comparisons. - **State of AI 2024 Report sparks insights**: The **State of AI Report 2024** by **Nathan Benaich** discusses major trends and investments in AI, minimizing mentions of specific models like *Torchtune*. - The report serves to inform and inspire discussions on AI's trajectory, particularly impact in fields like medicine and biology. - **LegoScale promises advances in distributed training**: **LegoScale** presents a customizable, PyTorch-native system enabling modular 3D parallel pre-training for large language models, enhancing performance significantly. - It aims to simplify complex training across multiple libraries and vast GPU ecosystems, potentially revolutionizing distributed training approaches. - **ICLR submissions available on OpenReview**: Members discuss how ICLR utilizes OpenReview to access preprints and track reviewer comments throughout the submission process. - This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.- Prompt Caching (beta) - Anthropic Aria at launch, highlighting a need for clearer head-to-head comparisons. - **State of AI 2024 Report sparks insights**: The **State of AI Report 2024** by **Nathan Benaich** discusses major trends and investments in AI, minimizing mentions of specific models like *Torchtune*. - The report serves to inform and inspire discussions on AI's trajectory, particularly impact in fields like medicine and biology. - **LegoScale promises advances in distributed training**: **LegoScale** presents a customizable, PyTorch-native system enabling modular 3D parallel pre-training for large language models, enhancing performance significantly. - It aims to simplify complex training across multiple libraries and vast GPU ecosystems, potentially revolutionizing distributed training approaches. - **ICLR submissions available on OpenReview**: Members discuss how ICLR utilizes OpenReview to access preprints and track reviewer comments throughout the submission process. - This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.
: no description founding a need for clearer head-to-head comparisons. - **State of AI 2024 Report sparks insights**: The **State of AI Report 2024** by **Nathan Benaich** discusses major trends and investments in AI, minimizing mentions of specific models like *Torchtune*. - The report serves to inform and inspire discussions on AI's trajectory, particularly impact in fields like medicine and biology. - **LegoScale promises advances in distributed training**: **LegoScale** presents a customizable, PyTorch-native system enabling modular 3D parallel pre-training for large language models, enhancing performance significantly. - It aims to simplify complex training across multiple libraries and vast GPU ecosystems, potentially revolutionizing distributed training approaches. - **ICLR submissions available on OpenReview**: Members discuss how ICLR utilizes OpenReview to access preprints and track reviewer comments throughout the submission process. - This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.- The Only Prompt - TheOnlyOSort sparks insights**: The **State of AI Report 2024** by **Nathan Benaich** discusses major trends and investments in AI, minimizing mentions of specific models like *Torchtune*. - The report serves to inform and inspire discussions on AI's trajectory, particularly impact in fields like medicine and biology. - **LegoScale promises advances in distributed training**: **LegoScale** presents a customizable, PyTorch-native system enabling modular 3D parallel pre-training for large language models, enhancing performance significantly. - It aims to simplify complex training across multiple libraries and vast GPU ecosystems, potentially revolutionizing distributed training approaches. - **ICLR submissions available on OpenReview**: Members discuss how ICLR utilizes OpenReview to access preprints and track reviewer comments throughout the submission process. - This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.
: The Only Prompt provides AI tools tailored to develop websites and web experiences, leveraging AI capabilities within its dedicated ecosystem.ecific models like *Torchtune*. - The report serves to inform and inspire discussions on AI's trajectory, particularly impact in fields like medicine and biology. - **LegoScale promises advances in distributed training**: **LegoScale** presents a customizable, PyTorch-native system enabling modular 3D parallel pre-training for large language models, enhancing performance significantly. - It aims to simplify complex training across multiple libraries and vast GPU ecosystems, potentially revolutionizing distributed training approaches. - **ICLR submissions available on OpenReview**: Members discuss how ICLR utilizes OpenReview to access preprints and track reviewer comments throughout the submission process. - This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.- GitHub - sogaiu/tree-sitter-clojure: Clojure(Script) grammar for tree-sitterke medicine and biology. - **LegoScale promises advances in distributed training**: **LegoScale** presents a customizable, PyTorch-native system enabling modular 3D parallel pre-training for large language models, enhancing performance significantly. - It aims to simplify complex training across multiple libraries and vast GPU ecosystems, potentially revolutionizing distributed training approaches. - **ICLR submissions available on OpenReview**: Members discuss how ICLR utilizes OpenReview to access preprints and track reviewer comments throughout the submission process. - This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.
: Clojure(Script) grammar for tree-sitter. Contribute to sogaiu/tree-sitter-clojure development by creating an account on GitHub.tive system enabling modular 3D parallel pre-training for large language models, enhancing performance significantly. - It aims to simplify complex training across multiple libraries and vast GPU ecosystems, potentially revolutionizing distributed training approaches. - **ICLR submissions available on OpenReview**: Members discuss how ICLR utilizes OpenReview to access preprints and track reviewer comments throughout the submission process. - This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.- LLM Leaderboard - Compare GPT-4o, Llama 3, Mistral, Gemini & other models | Artificial Analysisultiple libraries and vast GPU ecosystems, potentially revolutionizing distributed training approaches. - **ICLR submissions available on OpenReview**: Members discuss how ICLR utilizes OpenReview to access preprints and track reviewer comments throughout the submission process. - This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.
: Comparison and ranking the performance of over 30 AI models (LLMs) across key metrics including quality, price, performance and speed (output speed - tokens per second & latency - TTFT), context w...eprints and track reviewer comments throughout the submission process. - This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.- Llama 3.1 405B: API Provider Performance Benchmarking & Price Analysis | Artificial AnalysisnksMentioned">: Analysis of API providers for Llama 3.1 Instruct 405B across performance metrics including latency (time to first token), output speed (output tokens per second), price and others. API providers bench...nload here.
- Qwen2.5: A Party of Foundation Models!ixture-of-Experts Model: GITHUB HUGGING FACE MODELSCOPE DEMO DISCORD Introduction In the past three months since Qwen2’s release, numerous developers have built new models on the Qwen2 language models, providing us with... - **Bug Report on Shared Memory Resources**: The GitHub issue outlines a bug where using **flex attention** results in an out-of-resources error with shared memory. - It includes a minimal reproduction code snippet and encourages collaboration to resolve the identified problem. **Link mentioned**: - It includes a minimal reproduction code snippet and encourages collaboration to resolve the identified problem. **Link mentioned**: **Link mentioned**:
aider (Paul Gauthier) ▷ #questions-and-tips (67 messages🔥🔥):
Aider tutorials and resources
Aider context management
Aider API integration challenges
Usage of repo maps in Aider
Managing comments in code edits
- Tutorials and Resources for Aider: Users discovered a range of tutorial videos showcasing how to effectively use Aider for coding tasks.
- Highlighted videos include using Aider to build non-trivial apps and implementing features on Replit.
- Context Management Questions: Several members expressed concerns about how Aider handles context, specifically questioning if it sends modified files or previous commits to the chat.
- It was clarified that only committed versions are sent unless specified otherwise, raising the need for better communication of context by Aider.
- API Integration Challenges with Aider: Users faced issues with repeated API requests when the API provider had rate limits, seeking ways to prevent automatic retries that lead to freezing access.
- The community discussed effective ways to configure settings for custom API integrations to avoid blocking the service.
- Using Repo Maps for Code Context: Discussion on how to utilize the repo map feature in Aider to help the model retrieve and understand relevant code from the project.
- Users suggested creating effective library maps and preemptively adding files to the chat for Aider to offer better support in coding tasks.
- Managing Comments During Code Edits: Concerns were raised about Aider's tendency to strip comments from code even when users instructed it not to remove them.
- This sparked inquiries on methods to better preserve comments during automated edits and the reliability of existing prompts.
Links mentionedoned">:-
DevDocsof.ai/">: Fast, offline, and free documentation browser for developers. Search 100+ docs in one web app including HTML, CSS, JavaScript, PHP, Ruby, Python, Go, C, C++, and many more.05993">
- Sonnet seems as good as everdalities. Multimodal native AI models are essential to integrate real-world information and deliver comprehensive understanding. While proprietary multimodal native mode...
: Sonnet’s score on the aider code editing benchmark has been stable since it launched. deliver comprehensive understanding. While proprietary multimodal native mode... - FAQs://openreview.net/forum?id=SFN6Wm7YBI">: Frequently asked questions about aider.cale: One-stop PyTorch native solution for production ready...
- FAQ of large language models (LLMs) has been instrumental in advancing state-of-the-art natural language processing applications. Training LLMs with billions of parameters and trillions... : Frequently asked questions about aider.trumental in advancing state-of-the-art natural language processing applications. Training LLMs with billions of parameters and trillions...
- Repository mapplications. Training LLMs with billions of parameters and trillions... : Aider uses a map of your git repository to provide code context to LLMs.>
- Tips introduce Pixtral-12B, a 12--billion-parameter multimodal language model. Pixtral-12B is trained to understand both natural images and documents, achieving leading performance on various multimodal... : Tips for AI pair programming with aider.r multimodal language model. Pixtral-12B is trained to understand both natural images and documents, achieving leading performance on various multimodal...
- FAQhieving leading performance on various multimodal...
: Frequently asked questions about aider.odal...- Options references/1038097195422978059/1038097196224086148/1294440757809385504)** (15 messages🔥): >: Details about all of aider’s settings.94440757809385504)** (15 messages🔥): >
- Tutorial videos `Swarm.js Launch` >: Intro and tutorial videos made by aider users.t` >
- SECRET SAUCE of AI Coding? AI Devlog with Aider, Cursor, Bun and Notionced that the LangChain Discord community will close on **October 31, 2024**, to focus on building a new and improved community. Members are encouraged to fill out a form to stay updated about the changes. - Feedback is welcomed at community@langchain.dev, and there’s an invitation for anyone interested in becoming a moderator. - **Swarm.js Launch**: Pulkitgarg introduced **Swarm.js**, a Node.js SDK for orchestrating multi-agent systems using OpenAI's API, aimed at the JavaScript community. It includes features for defining agents with custom instructions, making the process simple and ergonomic. - Developers are encouraged to [check out the GitHub repo](https://github.com/youseai/openai-swarm-node) and contribute to the project. - **Real Estate LangChain Project**: Gustaf_81960_10487 shared their open-sourced LangChain real estate project that helped them secure a job, offering to aid contributors with letters of recommendation. They created a GitHub repo for the project and provided a link to their site, [Condo Cube](https://condo-cube.com/). - Contributors are encouraged to review the repo and get involved in solving real-world issues. - **OpenAI API Integration Inquiry**: Mikef0x inquired about plans for integrating OpenAI's real-time API, mentioning issues with Pydantic. Members discussed their experiences with Azure OpenAI and SQLDatabase, noting some humorous outcomes in responses. - **ImageMessage Type Discussion**: Grigorij0543 proposed introducing an **ImageMessage** type in LangChain to simplify image handling in chats. Clay_ferguson commented that text is usually associated with images and suggested the need for a more versatile file association method.
: What's the secret sauce of HIGH OUTPUT AI Coding?🔗 More AI Coding with AIDERhttps://youtu.be/ag-KxYS8Vuw🚀 More AI Coding with Cursorhttps://youtu.be/V9_Rzj...PHP search/replace issues
Troubleshooting Aider
LLM behavior
Model capability in Aider
- PHP Replacement Ruins Code Syntax: A member shared frustrations with PHP where search and replace operations lead to code duplication instead of proper replacements, resulting in lint errors.
- “...the code it wanted to replace” showcases the common issue of incomplete edits in PHP.
- Troubleshooting Page Offers Guidance: Another member suggested checking the troubleshooting page for issues when edits don’t apply to local files.
- This guidance included examples of messages like “Failed to apply edit to filename” as indicators of LLM disobedience.
- Model Strength Can Influence Outcomes: It's recommended to switch to stronger models like GPT-4o or Claude 3.5 Sonnet for better reliability with edits in Aider.
- Weaker models are noted to often disobey system prompt instructions, leading to common editing errors.
- Switching to Whole Format to Fix Edits: A member mentioned using the whole format approach to try to resolve frequent PHP editing issues.
- This indicates a proactive approach to mitigate errors when using Aider with PHP.
Nous Research Overview
DisTrO Project
Hermes 3 Model
Blockchain Development Opportunities
Community Engagement in AI
- Nous Research Overview: Nous Research, initially a Discord group, has evolved into a funded startup focused on AI research and development, especially in open-source projects.
- The community facilitates collaboration, sharing of ideas, and discussions related to AI research among its members.
- DisTrO Project: DisTrO is a project by Nous Research that enables faster AI model training across the public internet and emphasizes community-driven development.
- The project aims to provide an alternative to closed AI models, ensuring continued advancement in the open-source domain.
- Hermes 3 Model: Hermes 3 is an open-source fine-tuned language model that organizations use for various applications, showcased on platforms like lambda.chat.
- The model is part of the broader initiative to provide accessible AI tools without the restrictions typically associated with proprietary models.
- Blockchain Development Opportunities: Members discussed the availability of blockchain developers within the community, with offers for collaboration on related projects.
- Talent in blockchain development is welcomed as the integration of AI and blockchain technologies becomes increasingly relevant.
- Community Engagement in AI: The Discord channel serves as a hub for members to share links, insights, and to discuss their ongoing AI-related projects.
- Participants are encouraged to explore various AI tools, chat with language models, and foster connections to enhance their research efforts.
- Tweet from undefined/openai-swarm-node">: no description found - youseai/openai-swarm-node: Swarm.js is a Node.js implementation of OpenAI’s experimental Swarm framework. This SDK allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, while leveraging Node.js for building scalable, real-world applications.
- Tweet from Victor.crypto (@Victor_crypto_) allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, while leveraging Node.js for building scalable, real-world applications.: Ce projet IA est un ÉNORME BANGER @NousResearch est un collectif de chercheurs en IA open source, qui a récemment fait des avancées fracassantes pour l’IA décentralisée TL;DR : ➤ Ils ont entrainé ...implementation of OpenAI’s experimental Swarm framework. This SDK allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, w... --- ### **LangChain AI ▷ #[langchain-templates](https://discord.com/channels/1038097195422978059/1170025009960456282/1295025574913376318)** (1 messages): >
- arcee-ai/SuperNova-Medius · Hugging Facent systems using OpenAI’s API in a lightweight and ergonomic way, w...
---
### **LangChain AI ▷ #[langchain-templates](https://discord.com/channels/1038097195422978059/1170025009960456282/1295025574913376318)** (1 messages):
>: no description foundAPI in a lightweight and ergonomic way, w...
---
### **LangChain AI ▷ #[langchain-templates](https://discord.com/channels/1038097195422978059/1170025009960456282/1295025574913376318)** (1 messages):
>
- DisTrO and the Quest for Community-Trained AI Models | Andreessen Horowitz25574913376318)** (1 messages): >: Bowen Peng and Jeffrey Quesnelle of Nous Research discuss their mission to accelerate open source AI research, including with a new project called DisTrO. Launches for Node.js Enthusiasts**: Introducing [Swarm.js](https://github.com/youseai/openai-swarm-node), a lightweight Node.js SDK for orchestrating **multi-agent systems** using the **OpenAI API**. - This framework, inspired by the original Python version, allows developers to easily create and manage agents that perform tasks and communicate with each other. - **Key Features of Swarm.js Unveiled**: Swarm.js boasts features like **simple orchestration** of agents, custom instructions, and a lightweight structure designed for **real-world applications**. - Its ergonomic design aims to enhance usability and flexibility for developers looking to harness OpenAI's capabilities. - **Installation and Getting Started**: To install Swarm.js, simply run `npm install openai-swarm-node` and begin integrating with the OpenAI API. - The launch encourages both beginners and experts to jump in and explore the framework's potential in their projects. - **Community Contributions Are Welcome!**: The project lead invites users to contribute by checking out the [GitHub repo](https://github.com/youseai/openai-swarm-node), raising issues, and submitting pull requests. - This collaborative approach aims to enrich the SDK and foster a vibrant community around Swarm.js. **Link mentioned**:
- Facepalm Really GIF - Facepalm Really Stressed - Discover & Share GIFs systems** using the **OpenAI API**. - This framework, inspired by the original Python version, allows developers to easily create and manage agents that perform tasks and communicate with each other. - **Key Features of Swarm.js Unveiled**: Swarm.js boasts features like **simple orchestration** of agents, custom instructions, and a lightweight structure designed for **real-world applications**. - Its ergonomic design aims to enhance usability and flexibility for developers looking to harness OpenAI's capabilities. - **Installation and Getting Started**: To install Swarm.js, simply run `npm install openai-swarm-node` and begin integrating with the OpenAI API. - The launch encourages both beginners and experts to jump in and explore the framework's potential in their projects. - **Community Contributions Are Welcome!**: The project lead invites users to contribute by checking out the [GitHub repo](https://github.com/youseai/openai-swarm-node), raising issues, and submitting pull requests. - This collaborative approach aims to enrich the SDK and foster a vibrant community around Swarm.js. **Link mentioned**: : Click to view the GIFAI API**. - This framework, inspired by the original Python version, allows developers to easily create and manage agents that perform tasks and communicate with each other. - **Key Features of Swarm.js Unveiled**: Swarm.js boasts features like **simple orchestration** of agents, custom instructions, and a lightweight structure designed for **real-world applications**. - Its ergonomic design aims to enhance usability and flexibility for developers looking to harness OpenAI's capabilities. - **Installation and Getting Started**: To install Swarm.js, simply run `npm install openai-swarm-node` and begin integrating with the OpenAI API. - The launch encourages both beginners and experts to jump in and explore the framework's potential in their projects. - **Community Contributions Are Welcome!**: The project lead invites users to contribute by checking out the [GitHub repo](https://github.com/youseai/openai-swarm-node), raising issues, and submitting pull requests. - This collaborative approach aims to enrich the SDK and foster a vibrant community around Swarm.js. **Link mentioned**:
- Tweet from nano (@nanulled)ily create and manage agents that perform tasks and communicate with each other. - **Key Features of Swarm.js Unveiled**: Swarm.js boasts features like **simple orchestration** of agents, custom instructions, and a lightweight structure designed for **real-world applications**. - Its ergonomic design aims to enhance usability and flexibility for developers looking to harness OpenAI's capabilities. - **Installation and Getting Started**: To install Swarm.js, simply run `npm install openai-swarm-node` and begin integrating with the OpenAI API. - The launch encourages both beginners and experts to jump in and explore the framework's potential in their projects. - **Community Contributions Are Welcome!**: The project lead invites users to contribute by checking out the [GitHub repo](https://github.com/youseai/openai-swarm-node), raising issues, and submitting pull requests. - This collaborative approach aims to enrich the SDK and foster a vibrant community around Swarm.js. **Link mentioned**: : Here is a Test-time compute in action model: Qwen-2.5-0.5B sampler: entropix-local the continuous chain helps the model to recover and reach the definite answer the screenshot is a small slice of big ...and a lightweight structure designed for **real-world applications**. - Its ergonomic design aims to enhance usability and flexibility for developers looking to harness OpenAI's capabilities. - **Installation and Getting Started**: To install Swarm.js, simply run `npm install openai-swarm-node` and begin integrating with the OpenAI API. - The launch encourages both beginners and experts to jump in and explore the framework's potential in their projects. - **Community Contributions Are Welcome!**: The project lead invites users to contribute by checking out the [GitHub repo](https://github.com/youseai/openai-swarm-node), raising issues, and submitting pull requests. - This collaborative approach aims to enrich the SDK and foster a vibrant community around Swarm.js. **Link mentioned**:
- Tweet from Nick Dobos (@NickADobos)hance usability and flexibility for developers looking to harness OpenAI's capabilities. - **Installation and Getting Started**: To install Swarm.js, simply run `npm install openai-swarm-node` and begin integrating with the OpenAI API. - The launch encourages both beginners and experts to jump in and explore the framework's potential in their projects. - **Community Contributions Are Welcome!**: The project lead invites users to contribute by checking out the [GitHub repo](https://github.com/youseai/openai-swarm-node), raising issues, and submitting pull requests. - This collaborative approach aims to enrich the SDK and foster a vibrant community around Swarm.js. **Link mentioned**: : Grok api finally available! now on open router!!!o harness OpenAI's capabilities. - **Installation and Getting Started**: To install Swarm.js, simply run `npm install openai-swarm-node` and begin integrating with the OpenAI API. - The launch encourages both beginners and experts to jump in and explore the framework's potential in their projects. - **Community Contributions Are Welcome!**: The project lead invites users to contribute by checking out the [GitHub repo](https://github.com/youseai/openai-swarm-node), raising issues, and submitting pull requests. - This collaborative approach aims to enrich the SDK and foster a vibrant community around Swarm.js. **Link mentioned**:
- Tweet from samsja (@samsja19)rm-node` and begin integrating with the OpenAI API. - The launch encourages both beginners and experts to jump in and explore the framework's potential in their projects. - **Community Contributions Are Welcome!**: The project lead invites users to contribute by checking out the [GitHub repo](https://github.com/youseai/openai-swarm-node), raising issues, and submitting pull requests. - This collaborative approach aims to enrich the SDK and foster a vibrant community around Swarm.js. **Link mentioned**: : How are we training a model across datacenter ? We have more than 4 data center connected together. Yet we only communicate 1 minutes every 45 minutes for a 10b model. Quoting Prime Intellect (@Prim...come!**: The project lead invites users to contribute by checking out the [GitHub repo](https://github.com/youseai/openai-swarm-node), raising issues, and submitting pull requests. - This collaborative approach aims to enrich the SDK and foster a vibrant community around Swarm.js. **Link mentioned**:
- Jashancutie Bumsekichu GIF - JashanCutie BumSeKichu - Discover & Share GIFsl requests. - This collaborative approach aims to enrich the SDK and foster a vibrant community around Swarm.js. **Link mentioned**: : Click to view the GIFborative approach aims to enrich the SDK and foster a vibrant community around Swarm.js. **Link mentioned**:
- - YouTube Swarm.js. **Link mentioned**: : no description foundoned**:
- Tweet from Concerned Citizen (@BGatesIsaPyscho)imental Swarm framework. This SDK allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, while leveraging Node.js for building scalable, real-world applications.: Chinese New Police Robot. Pure Dystopia.pers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, while leveraging Node.js for building scalable, real-world applications.
- Tweet from Mihai Chirculescu (@m_chirculescu)orld applications.: 🚀 Launching backtrack_sampler - for experimenting with custom LLM sampling strategies that can rollback and regenerate tokens. It works with both @ggerganov 's llama.cpp (GGUF files) and @huggin...onomic way, w... --- ### **LangChain AI ▷ #[share-your-work](https://discord.com/channels/1038097195422978059/1038097372695236729/1294671840668815421)** (5 messages): >
- UTF-8 - Wikipedia//discord.com/channels/1038097195422978059/1038097372695236729/1294671840668815421)** (5 messages): >: no description found8097195422978059/1038097372695236729/1294671840668815421)** (5 messages): >
- DisTrO/A_Preliminary_Report_on_DisTrO.pdf at main · NousResearch/DisTrOs SDK` - **bootstrap-rag v0.0.9 is live!**: The early release of [bootstrap-rag v0.0.9](https://pypi.org/project/bootstrap-rag/) introduces removal of unwanted test code, along with bug fixes and documentation improvements. - Key features include LangChain integration, MLflow-evals support, and tested Qdrant templates, enhancing RAG capabilities. - **Exploring Contextual Retrieval**: A new [YouTube video](https://www.youtube.com/watch?v=n3nKTw83LW4) titled 'Contextual Retrieval using Langchain and OpenAI Swarm Agent' guides viewers through implementing contextual retrieval. - The video promises insights into using Langchain with OpenAI’s Swarm Agent for effective information retrieval. - **Introducing Swarm.js for Node.js**: [Swarm.js](https://github.com/youseai/openai-swarm-node) is launched as a lightweight Node.js SDK for orchestrating multi-agent systems utilizing OpenAI's API, designed for the JavaScript community. - This framework features simple multi-agent orchestration and is customizable, inviting contributions from developers at all skill levels.
: Distributed Training Over-The-Internet. Contribute to NousResearch/DisTrO development by creating an account on GitHub. **Link mentioned**: **Link mentioned**: --- ### **LangChain AI ▷ #[langchain-templates](https://discord.com/channels/1038097195422978059/1170025009960456282/1295025574913376318)** (1 messages): >
Nous Research AI ▷ #ask-about-llms (5 messages):
Best model for RP
Fine-tuning Llama 3.2
Unsloth Tool
- Discussion on Best Model for RP: A member inquired about the current best model for RP, signaling interest in general model performance.
- No specific model was mentioned in the conversation following this question.
- Fine-tuning Llama 3.2 Techniques: Another member, @deckard.1968, requested assistance with fine-tuning Llama 3.2, having prepared a JSON file with 800 input-output pairs.
- This request for help indicates active engagement in model enhancement within the community.
- Using Unsloth for Efficient Tuning: A suggestion was made to check out unsloth as a potential tool for fine-tuning models like Llama 3.2, noting its efficiency.
- Members pointed to the GitHub link for unsloth which claims to fine-tune various LLMs 2-5x faster with 80% less memory.
Nous Research AI ▷ #research-papers (6 messages):
ArXiv Performance
Model Collapse Phenomenon
MMedAgent Framework
GSM-Symbolic Benchmark
Medical AI Research Updates
- ArXiv Performance Under Scrutiny: Members discussed the recent sluggishness of the ArXiv platform, with one noting it has 'been breaking all week'.
- Another member reported that it has been functioning fine, highlighting varying user experiences.
- Understanding Model Collapse in Neural Networks: A new paper introduced by azure2089 highlights the phenomenon of model collapse in large neural networks, indicating that even 1% synthetic data can lead to critical performance degradation.
- The research suggests that increasing model size may actually exacerbate model collapse, challenging current training paradigms.
- MMedAgent: Multi-modal Medical Tool Utilization: The leading paper of the week, titled MMedAgent, focuses on teaching a multi-modal agent to effectively use medical tools in patient care, as reported by OpenlifesciAI.
- The paper's findings were discussed, alongside applications of various models like ONCOPILOT and PharmacyGPT.
- GSM-Symbolic Benchmark for Model Assessment: A recent paper presented the GSM-Symbolic benchmark, which leverages symbolic templates to create diverse mathematical questions for assessing LLM performance.
- This new benchmark aims to provide more reliable metrics for evaluating the formal reasoning capabilities of LLMs.
Links mentioned:- Tweet from Open Life Science AI (@OpenlifesciAI): Last Week in Medical AI: Top Research Papers/Models 🏅 (October 5 - October 12, 2024) 🏅 Medical AI Paper of the Week: MMedAgent: Learning to Use Medical Tools with Multi-modal Agent Authors: (@Haoy...eai/openai-swarm-node">
- GSM-Symbolic: Understanding the Limitations of Mathematical Reasoning in Large Language Models allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, while leveraging Node.js for building scalable, real-world applications.: Recent advancements in Large Language Models (LLMs) have sparked interest in their formal reasoning capabilities, particularly in mathematics. The GSM8K benchmark is widely used to assess the mathemat...implementation of OpenAI’s experimental Swarm framework. This SDK allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, w...
- Strong Model Collapsedevelopers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, w... : Within the scaling laws paradigm, which underpins the training of large neural networks like ChatGPT and Llama, we consider a supervised regression setting and establish the existance of a strong form...
Nous Research AI ▷ #interesting-links (11 messages🔥):
Machines of Loving Grace
OpenAI Swarm Controversy
Market Response to Multi-Agent Frameworks
GPTSwarm Overview
- Machines of Loving Grace explores AI optimism: In a recent essay, the CEO of Anthropic outlined his vision for how AI could significantly improve the world, emphasizing the underestimated benefits it offers alongside its risks.
- The future is fundamentally positive for AI, according to him, if risks are managed appropriately.
- OpenAI faces accusations over Swarm code: A member criticized OpenAI for allegedly stealing from Kye Gomez's repository, claiming the Swarms framework was infringed upon.
- The accusation included plans to seek legal repercussions against OpenAI unless investments were made in their project.
- Discussion on multi-agent framework similarities: Members discussed that several multi-agent frameworks, including CrewAI and Langchain, share common concepts, suggesting this isn't unique to OpenAI.
- One member specifically noted that Kye's approach seemed like a marketing stunt, given the broader context of existing frameworks.
- Emergence of GPTSwarm platform: The GPTSwarm project aims to unify various prompt engineering techniques into a cohesive framework for LLM agents, streamlining agent collaboration.
- Their method involves using computational graphs to optimize both node-level prompts and agent orchestration, showcasing the versatility of the platform.
Links mentionedried instruction data during pretraining --- ### **OpenAccess AI Collective (axolotl) ▷ #[general-help](https://discord.com/channels/1104757954588196865/1110594519226925137/1294388853007843339)** (9 messages🔥): >:-
GPTSwarmolotl) ▷ #[general-help](https://discord.com/channels/1104757954588196865/1110594519226925137/1294388853007843339)** (9 messages🔥):
>: no description foundttps://discord.com/channels/1104757954588196865/1110594519226925137/1294388853007843339)** (9 messages🔥):
>
- Tweet from Ky⨋ Gom⨋z (U/ACC) (HIRING) (@KyeGomezB)sharing` >: The Swarms framework was the first ever production-grade multi-agent orchestration framework. OpenAI has stolen our name, code, and methodology. Everything down from the syntax of the agent structure ...e config from @palebluedot_32091, who confirmed its accuracy but mentioned potential improvements. - *At glance it looks correct*, said @le_mess, suggesting enabling **sample packing** in the config. - **Importance of sample packing**: Member @le_mess highlighted the need to enable **sample packing** while running the training command `accelerate launch -m axolotl.cli.train config.yml`. - They also mentioned challenges with **multi-GPU** setups and using **LoRA** in earlier discussions. - **Fine-tuning with existing adapter**: Member @vortexwault inquired about refining a fine-tuned model using **Llama 3** as a base with an adapter for better accuracy. - @palebluedot_32091 advised merging the adapter into a new base model before training on additional data. - **Guide for merging adapters**: When @vortexwault asked for guidance on merging the adapter, @palebluedot_32091 directed them to a [GitHub guide](https://github.com/axolotl-ai-cloud/axolotl?tab=readme-ov-file#merge-lora-to-base). - He also indicated that the new training run's config should include the line `base_model: '/path/to/merged/model'`. - **Follow-up on guidance**: @vortexwault thanked @palebluedot_32091 for the guidance and mentioned he would look into the recommended resource. - "Let me check it brother thank you," said @vortexwault, indicating his intent to follow the advice.
- Dario Amodei — Machines of Loving Grace glance it looks correct*, said @le_mess, suggesting enabling **sample packing** in the config. - **Importance of sample packing**: Member @le_mess highlighted the need to enable **sample packing** while running the training command `accelerate launch -m axolotl.cli.train config.yml`. - They also mentioned challenges with **multi-GPU** setups and using **LoRA** in earlier discussions. - **Fine-tuning with existing adapter**: Member @vortexwault inquired about refining a fine-tuned model using **Llama 3** as a base with an adapter for better accuracy. - @palebluedot_32091 advised merging the adapter into a new base model before training on additional data. - **Guide for merging adapters**: When @vortexwault asked for guidance on merging the adapter, @palebluedot_32091 directed them to a [GitHub guide](https://github.com/axolotl-ai-cloud/axolotl?tab=readme-ov-file#merge-lora-to-base). - He also indicated that the new training run's config should include the line `base_model: '/path/to/merged/model'`. - **Follow-up on guidance**: @vortexwault thanked @palebluedot_32091 for the guidance and mentioned he would look into the recommended resource. - "Let me check it brother thank you," said @vortexwault, indicating his intent to follow the advice.
: How AI Could Transform the World for the Betterenabling **sample packing** in the config. - **Importance of sample packing**: Member @le_mess highlighted the need to enable **sample packing** while running the training command `accelerate launch -m axolotl.cli.train config.yml`. - They also mentioned challenges with **multi-GPU** setups and using **LoRA** in earlier discussions. - **Fine-tuning with existing adapter**: Member @vortexwault inquired about refining a fine-tuned model using **Llama 3** as a base with an adapter for better accuracy. - @palebluedot_32091 advised merging the adapter into a new base model before training on additional data. - **Guide for merging adapters**: When @vortexwault asked for guidance on merging the adapter, @palebluedot_32091 directed them to a [GitHub guide](https://github.com/axolotl-ai-cloud/axolotl?tab=readme-ov-file#merge-lora-to-base). - He also indicated that the new training run's config should include the line `base_model: '/path/to/merged/model'`. - **Follow-up on guidance**: @vortexwault thanked @palebluedot_32091 for the guidance and mentioned he would look into the recommended resource. - "Let me check it brother thank you," said @vortexwault, indicating his intent to follow the advice.- GitHub - openai/swarm: Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.ate launch -m axolotl.cli.train config.yml`. - They also mentioned challenges with **multi-GPU** setups and using **LoRA** in earlier discussions. - **Fine-tuning with existing adapter**: Member @vortexwault inquired about refining a fine-tuned model using **Llama 3** as a base with an adapter for better accuracy. - @palebluedot_32091 advised merging the adapter into a new base model before training on additional data. - **Guide for merging adapters**: When @vortexwault asked for guidance on merging the adapter, @palebluedot_32091 directed them to a [GitHub guide](https://github.com/axolotl-ai-cloud/axolotl?tab=readme-ov-file#merge-lora-to-base). - He also indicated that the new training run's config should include the line `base_model: '/path/to/merged/model'`. - **Follow-up on guidance**: @vortexwault thanked @palebluedot_32091 for the guidance and mentioned he would look into the recommended resource. - "Let me check it brother thank you," said @vortexwault, indicating his intent to follow the advice.
: Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team. - openai/swarm ### **OpenAccess AI Collective (axolotl) ▷ #[general-help](https://discord.com/channels/1104757954588196865/1110594519226925137/1294388853007843339)** (9 messages🔥): > ### **OpenAccess AI Collective (axolotl) ▷ #[general-help](https://discord.com/channels/1104757954588196865/1110594519226925137/1294388853007843339)** (9 messages🔥): > >
Nous Research AI ▷ #research-papers (6 messages):
ArXiv performance issues
Model collapse in neural networks
Medical AI advancements
GSM-Symbolic benchmark for LLMs
Recent medical AI papers
- ArXiv performance issues reported: Users have expressed concerns about ArXiv being slow lately, with discussions highlighting that it has been 'breaking all week'.
- Others confirmed that they are experiencing normal performance, indicating mixed user experiences.
- Model collapse phenomenon in large neural networks: A study introduced the concept of model collapse, showing that even 1% synthetic data in training can degrade performance significantly.
- It was theorized that larger models trained under the scaling laws paradigm may actually worsen this collapse rather than improve it.
- GSM-Symbolic benchmark enhances LLM evaluations: A paper discussed the introduction of GSM-Symbolic, an improved benchmark for evaluating mathematical reasoning in LLMs, addressing issues with existing benchmarks.
- This new benchmark provides more reliable performance metrics through sectioned, controllable evaluations.
- Top Medical AI Papers Aid Healthcare: Several papers highlighted advancements in Medical AI, including MMedAgent, which focuses on learning to use medical tools through multi-modal approaches.
- The list features frameworks for rare disease phenotyping and a benchmark for evaluating medical LLM safety, among others.
- Discussion on various AI research papers: Users shared links and summarized multiple significant research papers, outlining innovative applications in AI related to medical diagnoses and genetic experiments.
- Discussions also centered around improving LLM factuality and the implications of using AI in healthcare ethics.
Links mentionedpenAccess AI Collective (axolotl) ▷ #[axolotl-help-bot](https://discord.com/channels/1104757954588196865/1225300056442409040/1295036626795495517)** (5 messages): >:-
Tweet from Open Life Science AI (@OpenlifesciAI)6865/1225300056442409040/1295036626795495517)** (5 messages):
>: Last Week in Medical AI: Top Research Papers/Models 🏅 (October 5 - October 12, 2024) 🏅 Medical AI Paper of the Week: MMedAgent: Learning to Use Medical Tools with Multi-modal Agent Authors: (@Haoy...th text completion training**: Training an instruct model, such as **GPT-3.5-Instruct**, on a text completion task can significantly improve its performance on instruction-based tasks, resulting in more accurate and contextually relevant completions.
- *“This can result in improved ability to generate text that aligns closely with prompts.”*
- **Risk of overfitting in narrow training**: There are potential risks of **overfitting** if the training dataset for the text completion task is not diverse or is too small, leading to high performance on familiar data but poor generalization.
- A caution emphasized was that **performance might degrade** on other tasks not seen during the specific training.
- **Adaptation to new formats and fields**: The training can help the model adapt to new formats or types of instructions that it was not exposed to during its initial training.
- This adaptation fosters the model's ability to handle a broader range of inputs more effectively.
- **Importance of diverse training datasets**: To achieve optimal outcomes when training instruct models, using a diverse and representative dataset is essential.
- Monitoring for overfitting and ensuring broad performance evaluation across tasks is crucial for maintaining generalization capabilities.
**Link mentioned**:
- GSM-Symbolic: Understanding the Limitations of Mathematical Reasoning in Large Language Modelserformance on instruction-based tasks, resulting in more accurate and contextually relevant completions. - *“This can result in improved ability to generate text that aligns closely with prompts.”* - **Risk of overfitting in narrow training**: There are potential risks of **overfitting** if the training dataset for the text completion task is not diverse or is too small, leading to high performance on familiar data but poor generalization. - A caution emphasized was that **performance might degrade** on other tasks not seen during the specific training. - **Adaptation to new formats and fields**: The training can help the model adapt to new formats or types of instructions that it was not exposed to during its initial training. - This adaptation fosters the model's ability to handle a broader range of inputs more effectively. - **Importance of diverse training datasets**: To achieve optimal outcomes when training instruct models, using a diverse and representative dataset is essential. - Monitoring for overfitting and ensuring broad performance evaluation across tasks is crucial for maintaining generalization capabilities. **Link mentioned**: : Recent advancements in Large Language Models (LLMs) have sparked interest in their formal reasoning capabilities, particularly in mathematics. The GSM8K benchmark is widely used to assess the mathemat... of overfitting in narrow training**: There are potential risks of **overfitting** if the training dataset for the text completion task is not diverse or is too small, leading to high performance on familiar data but poor generalization. - A caution emphasized was that **performance might degrade** on other tasks not seen during the specific training. - **Adaptation to new formats and fields**: The training can help the model adapt to new formats or types of instructions that it was not exposed to during its initial training. - This adaptation fosters the model's ability to handle a broader range of inputs more effectively. - **Importance of diverse training datasets**: To achieve optimal outcomes when training instruct models, using a diverse and representative dataset is essential. - Monitoring for overfitting and ensuring broad performance evaluation across tasks is crucial for maintaining generalization capabilities. **Link mentioned**:
- Strong Model Collapsefitting** if the training dataset for the text completion task is not diverse or is too small, leading to high performance on familiar data but poor generalization. - A caution emphasized was that **performance might degrade** on other tasks not seen during the specific training. - **Adaptation to new formats and fields**: The training can help the model adapt to new formats or types of instructions that it was not exposed to during its initial training. - This adaptation fosters the model's ability to handle a broader range of inputs more effectively. - **Importance of diverse training datasets**: To achieve optimal outcomes when training instruct models, using a diverse and representative dataset is essential. - Monitoring for overfitting and ensuring broad performance evaluation across tasks is crucial for maintaining generalization capabilities. **Link mentioned**: : Within the scaling laws paradigm, which underpins the training of large neural networks like ChatGPT and Llama, we consider a supervised regression setting and establish the existance of a strong form... ### **OpenAccess AI Collective (axolotl) ▷ #[axolotl-help-bot](https://discord.com/channels/1104757954588196865/1225300056442409040/1295036626795495517)** (5 messages): > ### **OpenAccess AI Collective (axolotl) ▷ #[axolotl-help-bot](https://discord.com/channels/1104757954588196865/1225300056442409040/1295036626795495517)** (5 messages): > >
Perplexity AI ▷ #announcements (1 messages):
Reasoning Mode
Pro Search features
Use cases for Reasoning
- *Reasoning Mode* Rolls Out in Pro Search: The Perplexity Team announced the launch of a new experimental feature called Reasoning Mode which detects when extra compute or searches can yield a better answer.
- Users are encouraged to explore this feature and share interesting use cases in designated feedback channel.
- *Pro Search Examples* to Try Out: Examples provided include inquiries like finding co-founders of OpenAI, listing top cryptocurrencies, and assessing the best rated films on IMDB.
- This feature also suggests gathering detailed information about DOW company CEOs including their LinkedIn URLs and tenure details.
Perplexity AI ▷ #general (205 messages🔥🔥):
Perplexity AI performance issues
Image generation with Perplexity
Reasoning feature in Pro Search
User experiences with various models
Coding capabilities of AI models
- Perplexity AI struggles with memory: Users report that Perplexity AI is having issues retaining conversational context, leading to frustration as it fails to recall previous messages during interactions.
- Some users speculate this may be related to API limitations or token issues, affecting model performance inconsistently across longer threads.
- Image generation capabilities discussed: There are inquiries about how to generate images using Perplexity AI, with references to guides available on Discord.
- Users were directed to check the web version for image generation options, but faced confusion over functionality across different platforms.
- Confusion over Reasoning feature: The newly introduced Reasoning feature in the Pro Search option has users questioning its mechanics, particularly regarding when and how it activates.
- Some users find the automatic activation of reasoning during searches either beneficial or detrimental, with varying results in response quality.
- Mixed user experiences with AI models: Users share mixed results using different AI models, with some indicating Claude performs better than others in coding tasks and interactions.
- Reports suggest that updates have impacted the performance of certain models like O1 mini, with users expressing concerns over its output accuracy.
- Seeking targeted search results: A user expresses frustration over the inability to filter search results to only retrieve information from a specific website using Perplexity.
- Despite using the 'site:' parameter, the results still included information from other sources, leaving users looking for a better solution for targeted searches.
Links mentionedith Aria` >:-
Tweet from Beff – e/acc (@BasedBeffJezos)er.com/rhymes_ai_), is now ranked first on the 🤗 Open LLM Leaderboard, featuring **24.9B parameters** and capable of handling image, video, and text inputs.
- It boasts a **64k token context window** and was pre-trained on a diverse dataset, including **400B multimodal tokens**, ensuring stable performance across tasks ([Paper](https://hf.co/papers/2410.05993), [Blog](https://rhymes.ai/blog-details/aria-first-open-multimodal-native-moe-model)).
- **Users Rave About Aria's Performance**: *This is the BEST vision language model I have ever tried!* claims a user regarding the **25.3B multimodal Aria model**, emphasizing its impressive capabilities with image and video inputs.
- The model has been released under the **Apache-2.0 license**, and fine-tuning scripts are also available to the community.
- **AI Reasoning Under Scrutiny**: A [YouTube video](https://www.youtube.com/watch?v=tTG_a0KPJAc) titled 'Apple DROPS AI BOMBSHELL: LLMS CANNOT Reason' explores critical discussions around the reasoning capabilities of language models.
- The creator encourages viewers to prepare for AGI, sharing links to their social media and related content, hinting at a broader conversation about AI's current limitations.
: Who's working on a Saturday? 👨💻 the 🤗 Open LLM Leaderboard, featuring **24.9B parameters** and capable of handling image, video, and text inputs. - It boasts a **64k token context window** and was pre-trained on a diverse dataset, including **400B multimodal tokens**, ensuring stable performance across tasks ([Paper](https://hf.co/papers/2410.05993), [Blog](https://rhymes.ai/blog-details/aria-first-open-multimodal-native-moe-model)). - **Users Rave About Aria's Performance**: *This is the BEST vision language model I have ever tried!* claims a user regarding the **25.3B multimodal Aria model**, emphasizing its impressive capabilities with image and video inputs. - The model has been released under the **Apache-2.0 license**, and fine-tuning scripts are also available to the community. - **AI Reasoning Under Scrutiny**: A [YouTube video](https://www.youtube.com/watch?v=tTG_a0KPJAc) titled 'Apple DROPS AI BOMBSHELL: LLMS CANNOT Reason' explores critical discussions around the reasoning capabilities of language models. - The creator encourages viewers to prepare for AGI, sharing links to their social media and related content, hinting at a broader conversation about AI's current limitations.
- Tweet from Phi Hoang (@apostraphi)uts. - It boasts a **64k token context window** and was pre-trained on a diverse dataset, including **400B multimodal tokens**, ensuring stable performance across tasks ([Paper](https://hf.co/papers/2410.05993), [Blog](https://rhymes.ai/blog-details/aria-first-open-multimodal-native-moe-model)). - **Users Rave About Aria's Performance**: *This is the BEST vision language model I have ever tried!* claims a user regarding the **25.3B multimodal Aria model**, emphasizing its impressive capabilities with image and video inputs. - The model has been released under the **Apache-2.0 license**, and fine-tuning scripts are also available to the community. - **AI Reasoning Under Scrutiny**: A [YouTube video](https://www.youtube.com/watch?v=tTG_a0KPJAc) titled 'Apple DROPS AI BOMBSHELL: LLMS CANNOT Reason' explores critical discussions around the reasoning capabilities of language models. - The creator encourages viewers to prepare for AGI, sharing links to their social media and related content, hinting at a broader conversation about AI's current limitations.
: edge of curiosity*64k token context window** and was pre-trained on a diverse dataset, including **400B multimodal tokens**, ensuring stable performance across tasks ([Paper](https://hf.co/papers/2410.05993), [Blog](https://rhymes.ai/blog-details/aria-first-open-multimodal-native-moe-model)). - **Users Rave About Aria's Performance**: *This is the BEST vision language model I have ever tried!* claims a user regarding the **25.3B multimodal Aria model**, emphasizing its impressive capabilities with image and video inputs. - The model has been released under the **Apache-2.0 license**, and fine-tuning scripts are also available to the community. - **AI Reasoning Under Scrutiny**: A [YouTube video](https://www.youtube.com/watch?v=tTG_a0KPJAc) titled 'Apple DROPS AI BOMBSHELL: LLMS CANNOT Reason' explores critical discussions around the reasoning capabilities of language models. - The creator encourages viewers to prepare for AGI, sharing links to their social media and related content, hinting at a broader conversation about AI's current limitations.- Per Watch **400B multimodal tokens**, ensuring stable performance across tasks ([Paper](https://hf.co/papers/2410.05993), [Blog](https://rhymes.ai/blog-details/aria-first-open-multimodal-native-moe-model)). - **Users Rave About Aria's Performance**: *This is the BEST vision language model I have ever tried!* claims a user regarding the **25.3B multimodal Aria model**, emphasizing its impressive capabilities with image and video inputs. - The model has been released under the **Apache-2.0 license**, and fine-tuning scripts are also available to the community. - **AI Reasoning Under Scrutiny**: A [YouTube video](https://www.youtube.com/watch?v=tTG_a0KPJAc) titled 'Apple DROPS AI BOMBSHELL: LLMS CANNOT Reason' explores critical discussions around the reasoning capabilities of language models. - The creator encourages viewers to prepare for AGI, sharing links to their social media and related content, hinting at a broader conversation about AI's current limitations.
: As a big fan of Perplexity AI, I built an Apple Watch App to have it just on my wrist. Per Watch is designed to provide fast, straightforward and efficient answers to users queries by using the powe... - It boasts a **64k token context window** and was pre-trained on a diverse dataset, including **400B multimodal tokens**, ensuring stable performance across tasks ([Paper](https://hf.co/papers/2410.05993), [Blog](https://rhymes.ai/blog-details/aria-first-open-multimodal-native-moe-model)). - **Users Rave About Aria's Performance**: *This is the BEST vision language model I have ever tried!* claims a user regarding the **25.3B multimodal Aria model**, emphasizing its impressive capabilities with image and video inputs. - The model has been released under the **Apache-2.0 license**, and fine-tuning scripts are also available to the community. - **AI Reasoning Under Scrutiny**: A [YouTube video](https://www.youtube.com/watch?v=tTG_a0KPJAc) titled 'Apple DROPS AI BOMBSHELL: LLMS CANNOT Reason' explores critical discussions around the reasoning capabilities of language models. - The creator encourages viewers to prepare for AGI, sharing links to their social media and related content, hinting at a broader conversation about AI's current limitations.- **Users Rave About Aria's Performance**: *This is the BEST vision language model I have ever tried!* claims a user regarding the **25.3B multimodal Aria model**, emphasizing its impressive capabilities with image and video inputs. - The model has been released under the **Apache-2.0 license**, and fine-tuning scripts are also available to the community. - **AI Reasoning Under Scrutiny**: A [YouTube video](https://www.youtube.com/watch?v=tTG_a0KPJAc) titled 'Apple DROPS AI BOMBSHELL: LLMS CANNOT Reason' explores critical discussions around the reasoning capabilities of language models. - The creator encourages viewers to prepare for AGI, sharing links to their social media and related content, hinting at a broader conversation about AI's current limitations.- **AI Reasoning Under Scrutiny**: A [YouTube video](https://www.youtube.com/watch?v=tTG_a0KPJAc) titled 'Apple DROPS AI BOMBSHELL: LLMS CANNOT Reason' explores critical discussions around the reasoning capabilities of language models. - The creator encourages viewers to prepare for AGI, sharing links to their social media and related content, hinting at a broader conversation about AI's current limitations.
Perplexity AI ▷ #sharing (19 messages🔥):
AI Image Generation Costs
Kardashev Scale of Civilization
Perplexity Pro Updates
SSD Booting from Ubuntu
AI for Critical Thinking
- Unlocking AI Image Generation Costs: A discussion emerged about the current costs associated with AI image generation and its implications for users seeking affordability. For more details, visit the link.
- It was noted that understanding these costs can help in budgeting for projects that involve sophisticated visual content.
- Exploring the Kardashev Scale: A member shared a YouTube video discussing the Kardashev Scale of Civilization Advancement, which measures a civilization's energy consumption. This scale offers a fascinating perspective on technological progress and energy utilization.
- The video sparked interest in its implications for the future of humanity and the advancement of technology.
- Perplexity Pro Features & Updates: Several messages highlighted discussions on the Perplexity Pro update, with users sharing their experiences and recommendations. This update seems to focus on enhancing user functionality and engagement.
- To dive deeper into updates, check out this link.
- Troubleshooting SSD Booting from Ubuntu: A user noted a situation where they needed guidance on booting into Windows from Ubuntu without accessing the grub menu. They mentioned a method involving SSH to run commands, highlighting an effective solution shared by the community.
- For a more extensive discussion, view the related link here.
- Critical Thinking Encouragement with AI: One member expressed a thought-provoking statement about how governments might prefer obedient workers over critically thinking individuals, alluding to Perplexity's Reasoning Beta feature. This feature aims to foster a larger population capable of critical thinking, emphasizing its importance in the information age.
- The community resonated with this perspective, underlining the need for tools that empower independent thought, as discussed in this link.
Mozilla AI ▷ #announcements (1 messages):
: no description found
Perplexity AI ▷ #pplx-api (7 messages):
API URL Responses
Perplexity API Support
Search Domain Filter
Requesting Similar Responses
Sonar Models
- API doesn't show source URLs: A user is seeking help on how to enable source URLs in responses from the API, noting that their recent email to support went unanswered.
- Hello anyone? expresses their frustration over the lack of responses.
- Inquiry about Perplexity's online API: A member questioned whether Perplexity supports an online API, prompting another user to reference
sonar-online
models available at Perplexity API Docs.- Details on various
llama-3.1-sonar
models were shared, including their parameter counts and context lengths for chat completion.
- Details on various
- Concerns with 'search_domain_filter' parameter: A user is testing the effectiveness of the 'search_domain_filter' parameter in API requests and questions its adherence to specific domains provided.
- Despite indicating a domain lacking the searched content, they received valid responses, raising doubts about the parameter's functionality.
- Requests for tips on API response similarity: A user is looking for tricks to achieve API responses similar to those of the online interface, noting a significant difference in response quality.
- They were uncertain if the API uses search capabilities while utilizing the model:
llama-3.1-sonar-large-128k-online
.
- They were uncertain if the API uses search capabilities while utilizing the model:
Link mentioned: Supported Models - Perplexity1271/)** (1 messages):
mihai4256: Starting from the above, I made this: https://github.com/Mihaiii/backtrack_sampler
Gorilla LLM (Berkeley Function Calling) ▷ #discussion (1 messages):
: no description found
GPU MODE ▷ #general (21 messages🔥):
Attention Layer using cuDNN
Profiling performance of PyTorch, Triton, and CUDA
Saturday talks recordings
CUDA documentation unicode issue
- Confusion over Attention Layer Implementation: A member seeks a tutorial or implementation for an Attention layer using cuDNN's SDPA in Python, expressing confusion about instantiating the pygraph.
- Any help would be appreciated as the member follows the notebook from the cudnn-frontend repository.
- Performance Profiling Discrepancies: A member profiles operations in PyTorch, Triton, and CUDA, finding significant differences in performance results.
- While Triton claims equal performance across the three, self-evaluation shows PyTorch leading, leading to questions about which profiler to trust.
- Saturday Talks Recording Availability: Inquiries are made about the recording of Saturday talks, specifically regarding the Metal talk, with confirmations that they are always recorded.
- A member looks for a link to the recorded talks, which is provided by another member.
- CUDA Documentation Unicode Confusion: A member points out that CUDA documentation improperly uses a unicode character in
#include
, instead of the standard/
.- Another member suggests this may relate to a rendering engine or font issue, inviting others to check the referenced documentation.
GPU MODE ▷ #triton (4 messages):
Using Triton for CUDA tasks
Bit packing techniques in Triton
Inline ASM and warp operations
- Triton versus pure CUDA for bit packing: A user expressed difficulties with Triton when trying to pack an Nx32 bool array into N 32-bit integers, suggesting that pure CUDA might be a better fit for such tasks.
- They mentioned that using
reduce
could accomplish the task, but noted that a single__ballot_sync
in CUDA offers better performance.
- They mentioned that using
- Reduce function not leveraging balloting: The user confirmed that their basic
reduce
implementation in Triton did not utilize balloting in the generatedptx
, which was expected but disappointing.- They shared a code snippet with a custom
bitwise_or
function, explaining their approach to combine boolean values.
- They shared a code snippet with a custom
- Challenges with inline ASM for bit packing: There was a query about the feasibility of using inline ASM for bit packing in Triton, with concerns about the lack of control over warp mapping.
- The user questioned whether inline ASM could solve their bit packing issues despite their unfamiliarity with Triton.
GPU MODE ▷ #torch (7 messages):
IMA errors in PyTorch CI
Learning Torch
Model Parallelism vs Data Parallelism
- Chaining IMA Errors in PyTorch CI Glitch: An issue was raised concerning IMA errors that appear sporadically in torchao CI following updates to the PyTorch nightly version, leading to numerous failing tests.
- It was suggested to isolate IMA errors using
CUDA_LAUNCH_BLOCKING=1
or leveragecompute-sanitizer --tool=memcheck
with caching disabled to surface the errors immediately.
- It was suggested to isolate IMA errors using
- Best Projects to Learn Torch: A simple project suggestion to learn Torch involves implementing traffic sign detection and classification, which serves as a practical starting point.
- After grasping the basics, one should explore advanced topics like DDP (Distributed Data Parallel) and FSDP (Fully Sharded Data Parallel) to deepen their understanding.
- Data Parallelism Comes First: The consensus noted that data parallelism should be the first approach to try due to its straightforward nature, with model parallelism considered only if memory constraints arise.
- It was emphasized that most users employing model parallelism would have already implemented data parallelism.
GPU MODE ▷ #announcements (1 messages):
Metal Programming
Apple M series chips
PyTorch kernel merging
- Kickoff to Metal Programming Session: We're starting in 25 minutes at noon PST for a session on Metal Programming tailored for beginners on Apple's M series chips.
- Join <@1247230890036166827> as they guide participants through writing and merging their first Metal kernel in PyTorch.
- Getting Ready for Metal Kernel Merging: The session will focus on helping attendees to write and merge their first Metal kernel.
- This is a great opportunity for those interested in diving into the world of Metal programming with hands-on guidance.
GPU MODE ▷ #algorithms (4 messages):
Entropix Sampling
Open Source Projects
LLM Samplers
- Skepticism Around Entropix Sampling: Discussion arose about Entropix sampling, with some expressing concerns about the credibility of proponents, leading to initial suspicions of it being a scam.
- Despite skepticism, there seems to be a promise in its approach that simplifies reasoning without extensive modifications.
- Difficulty in Understanding Entropix: Members shared frustration about finding explanations for Entropix that sounded like nonsensical cult stuff, highlighting the challenges in grasping its concept.
- This led to a quest for clearer, more concrete information that can demystify its workings.
- Insightful Blog Post on Entropix: A recent blog post discusses the buzz around Entropix, noting its aim to create o1-like reasoning using existing models with minimal modifications.
- The article elaborates on how it swaps out traditional samplers for algorithms based on entropy and varentropy without requiring re-training.
- Understanding LLM Samplers: The concept of a sampler was explained, emphasizing its role in predicting which word in a sequence comes next, based on calculated probabilities.
- The discussion revealed that there are various common approaches to sampling, pointing to the intricacies involved in developing effective LLM models.
Link mentioned: What is entropix doing? - Tim Kellogg: no description found
GPU MODE ▷ #cool-links (46 messages🔥):
INTELLECT-1 decentralized training
Diff Transformer paper
LegoScale framework
Communication methods in distributed training
OpenDiLoCo project
- INTELLECT-1: A Leap in Decentralized Training: Prime Intellect announced INTELLECT-1, heralded as the first-ever decentralized training of a 10B model, aiming to scale decentralized training by 10x.
- The project invites community involvement in building open-source AGI, sparking interest among members in distributed training methodologies.
- Exploring the Diff Transformer: The Diff Transformer paper presents a method to amplify attention to relevant context while canceling noise, enhancing performance in language modeling tasks.
- Members expressed curiosity regarding its effectiveness and the nature of the attention noise, with discussions suggesting a need for deeper proof of its claims.
- LegoScale: Modular 3D Parallelism for LLMs: A paper on LegoScale, a customizable, PyTorch-native system for 3D parallel pre-training of large language models, garnered excitement for its numerous innovations.
- It boasts features like customizable activation checkpointing and fp8 training, signaling a significant step forward in efficient model training.
- Communication Strategies in Distributed Training: Discussion on communication strategies during training highlighted the balance between efficiency and effectiveness in distributed setups, particularly concerning methods like diloco.
- Participants explored potential improvements in architecture and communication techniques, suggesting the possibility of using asymmetric neural net designs.
- Interest in OpenDiLoCo Project: Members showed enthusiasm for the OpenDiLoCo project, with discussions focused on its promising potential and recent developments within the decentralized training space.
- Notable mentions included community interactions with its team and interest in following its progression as a foundational element for upcoming innovations.
Links mentioned:- Tweet from Prime Intellect (@PrimeIntellect): Announcing INTELLECT-1: the first-ever decentralized training of a 10B model Scaling decentralized training 10x beyond prior efforts. Anyone can join us to build open-source AGI 🦋
- Differential Transformer: Transformer tends to overallocate attention to irrelevant context. In this work, we introduce Diff Transformer, which amplifies attention to the relevant context while canceling noise. Specifically, t...
- LegoScale: One-stop PyTorch native solution for production ready...: The development of large language models (LLMs) has been instrumental in advancing state-of-the-art natural language processing applications. Training LLMs with billions of parameters and trillions...
- GitHub - PrimeIntellect-ai/prime: prime (previously called ZeroBand) is a framework for efficient, globally distributed training of AI models over the internet.: prime (previously called ZeroBand) is a framework for efficient, globally distributed training of AI models over the internet. - PrimeIntellect-ai/prime
GPU MODE ▷ #beginner (34 messages🔥):
CUDA Programming on Windows
WSL and GPU Access
Dual Booting vs WSL2
Triton Kernel Debugging
Image Processing with GPUs
- CUDA Programming Options on Windows: Members discussed options for setting up CUDA programming on Windows, highlighting the use of WSL as a viable route, as well as dual booting for better performance.
- One noted that while 'Windows works, it gives fewer options' as compared to Linux alternatives.
- WSL GPU Access Confirmed: It's confirmed that WSL 2 can concurrently access the GPU with Windows, allowing for flexibility in CUDA development without needing a second GPU.
- Members pointed out the ease of making a second SSD bootable for a dual boot setup.
- Dual Booting Dilemmas: Concerns were raised about the challenges of dual booting, especially with gaming constraints on Linux, as members noted restrictions with games like those using Vanguard anti-cheat.
- One member humorously remarked, 'Vanguard is so invasive,' highlighting challenges faced with dual booting for gamers.
- Triton Kernel Debugging Tips: A user shared their experience debugging Triton kernels in VSCode, encountering issues with older versions overlapping with nightly installs.
- They figured out that installing the nightly version of Triton without removing the older version led to errors, specifically mentioning, 'name jit not defined'.
- Image Processing Insights: A newcomer sought advice on image processing using GPUs, indicating interest in engaging with the community on this topic.
- Their inquiry reflects the growing interest in leveraging GPU capabilities for image processing tasks.
Link mentioned: GPU MODE FAQ: FAQ What is GPU MODE? This is a community involved in high performance compute (mainly focusing on CUDA and other frameworks for ML/AI performance) Generally every week, a prominent engineer in the ...
GPU MODE ▷ #pmpp-book (1 messages):
rbatra: When will 5th edition come out?
GPU MODE ▷ #youtube-recordings (1 messages):
gau.nernst: https://youtu.be/cGtiaJjLkAI
GPU MODE ▷ #jax (2 messages):
Pallas kernels performance
JAX vs. PyTorch optimization
Custom operators for AI/ML
Float16 precision in LLM
Mistral-7B inference speed
- Pallas kernels struggle against JAX: Users express difficulties in finding examples where Pallas kernels outperform JAX, often noting that they can only match performance in best-case scenarios.
- Many examples highlight that the code is actually slower than the reference implementations, leading to skepticism about the speed claims for Pallas.
- Series on Custom Operator Optimization: A post discusses building custom operators to optimize AI/ML workloads, emphasizing the potential of Pallas in their implementation.
- The article is part of a three-part series aimed at maximizing TPU potential.
- Optimizing LLM Inference with Float16: An article details efforts to optimize Local LLM model inference speed in JAX using
float16
precision to match or surpass PyTorch performance.- The tests show that while the initial JAX implementation is slower, the optimized version can outperform existing PyTorch versions on RTX 3090.
Links mentioned:- The Rise of Pallas: Unlocking TPU Potential with Custom Kernels: Accelerating AI/ML Model Training with Custom Operators — Part 3
- Robert Dyro: no description found
GPU MODE ▷ #torchao (11 messages🔥):
cuBLAS restrictions
comfyui-flux-accelerator
Triton performance comparison
Acceleration of int8 ops
Gradient accumulation issue
- cuBLAS Path Restrictions Discussed: A member pointed out the restrictions for the cuBLAS path with
safe_int_mm
, referencing the pull request for a deeper understanding.- They observed that the restrictions look somewhat similar to those for FP8 operations.
- ComfyUI Flux Accelerator Cited: Discussion arose around the comfyui-flux-accelerator, noted for speeding up Flux.1 image generation, with members expressing interest in its utility.
- Another member confirmed it was not their work but still considered it a notable project.
- Triton Performance Boost on Linux: Triton is reported to be almost 2x faster on Linux as compared to vanilla Windows, indicating a significant performance upgrade.
- A member humorously noted they would implement improvements if they had the right skills.
- Accelerating Int8 and FP8 Operations Discussed: The conversation shifted towards the need for accelerated int8 ops on Amperes, emphasizing the importance of quantization.
- There was a recognition that if similar acceleration existed on Apple processors, it would greatly enhance performance.
- Gradient Accumulation and Offload Compatibility: A member questioned the compatibility of gradient accumulation with the CPUOffloadOptimizer set to offload=True, suggesting a workaround of accumulating on CPU.
- This raised discussion on the intricacies of grad handling in the optimization process.
Links mentioned:- GitHub - discus0434/comfyui-flux-accelerator: Accelerates Flux.1 image generation, just by using this node.: Accelerates Flux.1 image generation, just by using this node. - discus0434/comfyui-flux-accelerator
- Resubmit _int_mm by cpuhrsch · Pull Request #96685 · pytorch/pytorch: Avoids any changes to gemm_and_bias cc @soumith @voznesenskym @penguinwu @anijain2305 @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @Xia-Weiwen @wenzhe-nrv @jiayisunx @peterbe...
GPU MODE ▷ #sequence-parallel (1 messages):
navmarri.: do you mind tools you shared to make these nice animations?
GPU MODE ▷ #off-topic (6 messages):
Theta Wave Binaural Beats
Sleep Quality
Placebo Effect
Dreamlike Validation Prompt
- Trying Theta Wave Binaural Beats: One member shared their experience with theta wave binaural beats before sleep, emphasizing improved feelings on days with high sleep scores after prioritizing rest.
- They noted that while all-nighters may appeal to the younger crowd, achieving good sleep scores can be challenging for those with babies.
- Skepticism Around Efficacy: Another member expressed skepticism about the effectiveness of binaural beats, suggesting it might be snake oil.
- In response, a member pointed out that even if the claims are exaggerated, a placebo effect could still lead to positive experiences.
- Uncertain Impact of Binaural Beats: A participant shared anecdotal evidence that binaural beats do something, but the exact effects remain unclear and may not match existing research.
- This highlights a potential lack of clarity around the impact of binaural beats on sleep and relaxation.
- Concerns About Validation Prompt: A user expressed doubts about their choice of the word 'dreamlike' in the validation prompt, hinting at possible issues.
- This raises questions about the effectiveness of specific terminology in achieving desired user responses.
GPU MODE ▷ #llmdotc (1 messages):
madna11: Great approach
GPU MODE ▷ #rocm (13 messages🔥):
AMD Linux Radeon drivers
ROCm on alternative distros
hipBLASLt performance
- AMD updates Linux Radeon drivers: Finally, AMD made a big update to the Linux Radeon drivers to support the latest distros, including Ubuntu 24.04, although the HWE kernel is not supported yet. A member expressed frustration saying, 'My god why does Linux suck so much.'
- You can find more details about the update in the release notes.
- Choosing better distros for ROCm: Members discussed the advantages of using Arch Linux for ROCm as it provides better packaging and more cards have ROCm compiled out of the box. One shared their experience running ROCm on an RX6700XT with just commands for installation.
- Another member highlighted the potential for breakage in the Arch 6.0.2 version due to dependencies in the
extra
repository, while they personally moved to a distrobox Ubuntu container for 6.2.1.
- Another member highlighted the potential for breakage in the Arch 6.0.2 version due to dependencies in the
- Understanding hipBLASLt performance on MI300X: A question arose regarding the difference between rocBLAS and hipBLASLt, with the latter being noted as significantly faster on the MI300X. One member explained that 'hipBLASLt is a recent addition to the ecosystem, specifically optimized for tensor core usage on MI300X.'
- This suggests that users seeking performance improvements would benefit from using hipBLASLt with compatible hardware.
Link mentioned: Tweet from Osvaldo Doederlein (@opinali): Oh nice, AMD finally made a big update of the Linux Radeon drivers to support the latest distros, including my Ubuntu 24.04... but, not the HWE kernel yet. My god why does Linux suck so much. Anyway g...
GPU MODE ▷ #arm (8 messages🔥):
SIMD programming on ARM
Neon intrinsics
ARM Vectorization
Cross-Architecture Libraries
ARM Experimentation
- Exploring SIMD Programming Options on ARM: A member questioned the general consensus on getting into SIMD programming on ARM platforms like the RK3588 and the effectiveness of OpenCL versus SIMD intrinsics.
- Another member suggested that most developers use Neon compiler intrinsics from arm_neon.h or libraries with polyfills targeting various architectures.
- ATen Vectorized PR Improves Register Use: A member announced a PR that cleans up ATen Vectorized on ARM to utilize one vector register per instance instead of two.
- This change aims to optimize performance and efficiency in vectorized operations for ARM.
- Libraries vs Intrinsics for Portability: A participant highlighted that while intrinsics are effective, using libraries like vectorized or highway allows for easier portability across different architectures.
- These libraries help maintain efficiency without being tied to specific low-level implementations.
- Intrinsics with No Good AVX Analogues: It was mentioned that there are certain ARM instructions like FMLAL, a float16 FMA accumulating into float32, that lack direct AVX analogues.
- Optimizing ARM code requires unique considerations due to these specific instruction sets.
- M1 MacBook for ARM Experiments: One member observed that having an M1 MacBook negates the need to buy a Raspberry Pi for basic ARM experiments.
- This sentiment was quickly confirmed by another member, highlighting the efficiency of the M1 for learning purposes.
GPU MODE ▷ #liger-kernel (4 messages):
Loss Calculation in Training
Nanogpt Performance with Liger Kernel
- Clarification on Loss and Gradients: t_cc sought clarification on whether loss and gradients should be 0 where
label==ignore_index
and if to includeignore_index
in calculating batch mean loss.- qqs222 confirmed that this approach is similar to cross-entropy loss.
- Estimating Performance Boost with Liger Kernel: faresobeid inquired about the potential speed increase of a normal NanoGPT Pytorch model (124M) using Liger Kernel, trained on 3.5B tokens with a 64 batch size on 8xH100.
- However, specific estimates on performance improvements were not addressed in the conversation.
GPU MODE ▷ #metal (14 messages🔥):
Apple Silicon GPU with Docker
MPS Op Contributions
Kubernetes Device Plugin for Apple Silicon
- Apple Silicon GPUs and Docker: A discussion arose around the challenges of getting Docker to utilize Apple Silicon GPUs, with reports suggesting that this problem remains unsolved.
- One member mentioned that it has been an internal ticket for some time, yet no one is actively working on it, leading to a sense of hopelessness.
- Upcoming PR for PyTorch i0 Operator: A member announced the creation of a PR for torch.special.i0 during the session, indicating ongoing development efforts related to PyTorch MPS support.
- They are also working on the
bilinear2d_aa
operator, suggesting the next easy targets for contribution could be i1 and i0e.
- They are also working on the
- Kubernetes Needs Device Plugin for Apple GPUs: Another member raised a similar concern about using Apple Silicon GPUs within Kubernetes, advocating for a device plugin analogous to NVIDIA's version.
- This points to a broader need for ecosystem tools tailored for Apple's hardware in a containerized environment.
- MPS Operator Coverage Tracking: A member linked to an issue tracking MPS op coverage, which centralizes discussions on adding new operators for the MPS backend.
- They provided another link for reference which details the status of MPS ops support and contribution opportunities.
Links mentioned:- Apple Silicon GPUs, Docker and Ollama: Pick two.: If you've tried to use Ollama with Docker on an Apple GPU lately, you might find out that their GPU is not supported. But you can get Ollama to run with GPU support on a Mac. This article will ex...
- pytorc - Overview: pytorc has 2 repositories available. Follow their code on GitHub.
- GitHub - NVIDIA/k8s-device-plugin: NVIDIA device plugin for Kubernetes: NVIDIA device plugin for Kubernetes. Contribute to NVIDIA/k8s-device-plugin development by creating an account on GitHub.
- [MPS] Add i0 op by malfet · Pull Request #137849 · pytorch/pytorch: More-or-less verbatim copy of pytorch/aten/src/ATen/native/Math.h Line 101 in 47c8aa8 JITERATOR_HOST_DE...
- General MPS op coverage tracking issue · Issue #77764 · pytorch/pytorch: This issue is to have a centralized place to list and track work on adding support to new ops for the MPS backend. PyTorch MPS Ops Project : Project to track all the ops for MPS backend. There are ...
- MPS Support Matrix: no description found
GPU MODE ▷ #self-promotion (1 messages):
SGLang
FlashInfer
MLC LLM
LLM Deployment
Community Meetup
- Online Meetup for LLM Enthusiasts: Join the online meetup co-hosted by SGLang, FlashInfer, and MLC LLM to explore efficient LLM deployment and serving on October 5th at 4:00 PM PST. Register here for free!
- The agenda features insights on topics like low CPU overhead scheduling in SGLang and kernel generation for high-performance LLM serving.
- SGLang’s Insights on Efficient Scheduling: Liangsheng Yin, Lianmin Zheng, and Ke Bao will present updates on SGLang, focusing on deepseek MLA optimizations and fast JSON decoding during their session from 4:00 - 4:45 PM PST.
- A dedicated Q&A session will follow to clarify queries from the community.
- FlashInfer’s High-Performance Kernel Generation: From 4:50 - 5:35 PM PST, hear from Zihao Ye about the FlashInfer project, emphasizing techniques for kernel generation in performant LLM serving.
- Engage in discussions around advancements in this area with fellow developers.
- MLC LLM's Universal Deployment Strategies: Ruihang Lai, Yixin Dong, and Tianqi Chen will provide an overview of MLC LLM from 5:40 - 6:25 PM PST, discussing topics like low-latency serving and universal LLM deployment.
- Their session will also include a Q&A segment for participant interaction.
Link mentioned: LMSYS Online meetup on efficient LLM deployment and serving (Oct 16): We are excited to invite you to join the online meetup co-hosted by SGLang, FlashInfer, and MLC LLM! The three closely collaborating projects will share their different perspectives on efficient LLM d...
Cohere ▷ #discussions (121 messages🔥🔥):
CohereForAI contributions
Nobel Prize in Physics 2024
AI technology advancements
Hurricane impact on productivity
Community engagement in AI
- Exploring Contributions to CohereForAI: Members discussed the importance of meaningful contributions to the CohereForAI community, emphasizing citizen science as a point of entry for those interested in AI.
- One individual expressed a desire to actively contribute and mentor while also seeking projects that align with their vision of a symbiotic relationship with technology.
- Celebrating the Nobel Prize for AI Innovators: Sir John J. Hopfield and Sir Geoffrey E. Hinton were awarded the 2024 Nobel Prize in Physics for their groundbreaking work in AI and neural networks, as noted by a community member.
- Their research focuses on foundational discoveries enabling machine learning, showcasing the deep roots of AI in scientific endeavor.
- Aging Technology and Rapid Advancements: Community members reflected on the rapid pace of technological advancements, noting the importance of historical context from their personal experiences with computers growing up.
- Discussions acknowledged that while technology has progressed dramatically, its future remains uncertain and nuanced, with both potential upsides and conflicts inevitable.
- Impact of Hurricanes on Work Productivity: A member shared their situation regarding the ongoing hurricane affecting their ability to work, highlighting the broader community context.
- This led to discussions on how external events can significantly impact productivity and workflows in AI-related work.
- Desire for Dark Mode in Cohere's Interface: Members expressed a desire for a dark mode feature on the Cohere website, pointing out its utility for better user experience.
- The conversation emphasized the need for user-friendly enhancements, especially as the community interacts frequently with the platform.
Links mentioned:- The Nobel Prize in Physics 2024: The Nobel Prize in Physics 2024 was awarded jointly to John J. Hopfield and Geoffrey E. Hinton "for foundational discoveries and inventions that enable machine learning with artificial neural net...
- Cohere For AI (C4AI): Cohere For AI (C4AI) is Cohere's non-profit research lab that seeks to solve complex machine learning problems.
Cohere ▷ #questions (20 messages🔥):
Cohere AI Learning Contributions
Cohere Rerank Model Date Handling
Cohere Rerank Pricing
Command-R-08-2024 Model Obscurities
Prompt Engineering Best Practices
- Cohere AI contributions not clear: A user expressed concern about the usefulness of discussions on topics like consciousness if they do not contribute to the model's learning, stating their time is limited and they want meaningful engagement.
- They are seeking guidance on how to contribute effectively to machine learning while ensuring their insights are valued.
- Cohere Rerank model's date awareness: In relation to implementing Cohere Rerank, a user inquired whether the model inherently recognizes 'today' or if today's date needs to be provided for comparison with document dates.
- Clarity on this aspect could significantly impact how search results are prioritized based on recency.
- Confusion on Rerank pricing: A user questioned whether web-search pricing is included in the rerank pricing, as they couldn't find relevant information on the Cohere website.
- Understanding the pricing structure could help in planning for RAG implementation effectively.
- Command-R-08-2024 model's prompt issues: Users reported that when providing a second financial article to the Command-R-08-2024 model, it tends to analyze the first article instead of waiting for the new data.
- This indicates a possible limitation in how the model processes sequential user prompts, affecting its usability for untrained users.
- Effective prompt engineering tips: Discussion highlighted the importance of crafting concise prompts, noting that the Command-R-08-2024 performs better with shorter descriptions instead of longer instructions.
- Users are encouraged to utilize both the preamble and message parameters for clearer task directives, as indicated in the documentation.
Link mentioned: Crafting Effective Prompts (v1 API) — Cohere: This page describes different ways of crafting effective prompts for prompt engineering.
Cohere ▷ #api-discussions (7 messages):
API Tokens Usage
V2 API User Role Restrictions
Gen AI Hackathon Announcement
ChatBot Development Challenges
Cohere Code Sharing
- API Tokens not required for decent responses: A member questioned whether using
<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>
tokens is necessary for API requests, noting confusion around their importance.- Will the responses still be decent without including these tokens? remains unanswered but highlights a common concern among users.
- V2 API disallows user role after tool calls: A user inquired about the recent V2 API update which disallows the presence of a user role after a tool call in the sequence.
- This restriction seems to point towards a more streamlined interaction flow, but the reasoning behind it is unclear.
- Invitation to Gen AI Hackathon: An announcement detailed the Gen AI Hackathon organized by CreatorsCorner, encouraging collaboration amongst teams to create AI solutions.
- The challenge focuses on building multi-agent systems that are safe and enhance human potential through intelligent collaboration.
- ChatBot Development Approach: A member expressed interest in developing a chatbot but noted that it wouldn't answer every question, indicating specificity in its responses.
- This suggests a focus on tailored interactions rather than generic responses to user inquiries.
- Cohere Code Sharing Assistance: A member requested help in privately sharing the cohere code section, seeking assistance from others in the community.
- Feedback from another user suggested that employing a role system might have resolved the initial issue, emphasizing support among community members.
Link mentioned: Vertical Specific AI Agents Hackathon · Luma: Gen AI Agents CreatorsCorner, collaborating with Sambanova Systems, Prem, Marly, Senso and others enthusiastically welcomes individuals and teams to join our…
Cohere ▷ #projects (1 messages):
competent: Please don’t post that here
Modular (Mojo 🔥) ▷ #general (73 messages🔥🔥):
Mojo Backend Performance
Magic CLI Installation
Training Courses for Mojo
Issues with Mojo Playground
Community Support for Mojo
- Frustrations with Mojo Installation: A user expressed frustration over Mojo's installation issues, claiming that the playground is broken and that demo code leads to errors.
- They emphasized the lack of current tutorials and that they found the installation process for Magic, Mojo, and MAX disjointed and confusing.
- The Future of Mojo Performance: A user noted that while the language's syntax is promising, Mojo projects currently struggle with performance, especially when GPU support is lacking.
- They expressed a desire for working examples and a stable development environment to aid in learning and porting existing code.
- Magic CLI Confusion: Discussion arose about whether to use Magic with existing projects, with one user stating that they refuse to convert current code, fearing it might break functionality.
- The advice was given to either set up a
mojoproject.toml
file with Magic or to create a Conda environment for better management.
- The advice was given to either set up a
- Community Meeting Announcement: A user announced an upcoming community meeting on October 21st, inviting participants to add their topics to the agenda document.
- There was also an invitation to share Modular-related Discord server links to increase community engagement.
- Support and Resources for Mojo: The community encouraged sharing specific error messages and project details to improve support for Mojo users.
- Users were pointed towards documentation and resources for better understanding how to implement Magic and Mojo effectively.
Links mentioned:- Add MAX/Mojo to a conda project | Modular Docs: Although we recommend using Magic to manage your virtual environments
- Run inference with Python | Modular Docs: A walkthrough of the MAX Engine Python API, showing how to load and run a model.
- [Public] Modular Community Meeting: Modular Community Meeting This doc link: https://modul.ar/community-meeting-doc This is a public document; everybody is welcome to view and comment / suggest. All meeting participants must adhere to...
- Introducing Magic 🪄: Magic 🪄 is a package manager and virtual environment tool that unifies several otherwise separate tools into one, replacing our Modular CLI. In this video B...
Modular (Mojo 🔥) ▷ #mojo (66 messages🔥🔥):
AES hardware support in Mojo
Implicit conversions in Mojo Lists
Closed source compiler
Mojo library design
Mojo build issues
- AES hardware support implementation: A member shared their work on adding hardware AES support in Mojo through LLVM intrinsics, showing how intrinsics are dynamically invoked using string templating.
- They emphasized that this allows for tighter integration with LLVM even in library code, highlighting the flexibility of Mojo for such implementations.
- Implicit conversions in Mojo Lists: A member pointed out that Mojo Lists allow adding an Int to a List, questioning if this behavior is intentional due to a constructor for implicit conversions.
- It was explained that this behavior is incidental and that there is an ongoing issue tracking this potential inconsistency in the language.
- Mojo compiler is currently closed source: A member expressed interest in compiler development for Mojo but found out that the compiler is closed source to avoid design by committee.
- They were encouraged to investigate the open standard library for potential contributions, emphasizing that it allows for significant compiler-like design experiences.
- Building issues with Mojo: A user faced a linking error while trying to build Mojo files, indicating a missing library during the process.
- Help was offered, suggesting they verify if the magic environment was activated and if Mojo was properly installed via the command line tool.
- Success in binding Rust library to Mojo: A member successfully created a Mojo version of the Rust UUID library through FFI binding and published it on the prefix.dev platform.
- They encouraged others to refer to their project's build script for binding Rust libraries, sharing a GitHub link to their work for further reference.
Links mentioned:- Modular Docs: no description found
- GitHub - better-mojo/uuid: binding a mojo version of rust uuid.: binding a mojo version of rust uuid. Contribute to better-mojo/uuid development by creating an account on GitHub.
- [BUG]: Mojo permits the use of any constructor for implicit conversions · Issue #1310 · modularml/mojo: Bug description As title. Maybe it's working as intended, but I consider it a bug, for Mojo is not JavaScript. Is the behaviour there to circumvent the traitless-ness of print? Steps to reproduce ...
Modular (Mojo 🔥) ▷ #max (5 messages):
MAX installation on Linux
Compilation speed on MAX
Profiling compilation timing
Feature requests for MAX
- MAX Supports Multiple OS: MAX currently supports Linux, MacOS, and WSL2 on Windows, according to Modular Docs. This makes it accessible across different platforms for users starting out.
- Users are encouraged to install Magic for environment management using
curl -ssL https://magic.modular.com | bash
.
- Users are encouraged to install Magic for environment management using
- Compilation Times Need Attention: Compilation times are reported at around 300 ms for graphs after the first try, and 500 ms for the initial compilation, even for simple operations like tensor multiplication.
- Additionally, it's mentioned that a cache hit should be faster than the current observed times.
Links mentioned:- Get started with MAX | Modular Docs: On this page, we'll show you how to run some example projects.
- Issues · modularml/max: A collection of sample programs, notebooks, and tools which highlight the power of the MAX Platform - Issues · modularml/max
Latent Space ▷ #ai-general-chat (69 messages🔥🔥):
OpenAI Swarm
Entropix
RAG Techniques
LLM Reasoning
Jensen Interview Insights
- OpenAI Swarm emerges for multi-agent systems: OpenAI introduced Swarm, a lightweight library for building multi-agent systems, offering a stateless abstraction for managing interactions between agents.
- It provides insights into agent roles and handoffs while operating without using the Assistants API, with the creators emphasizing it remains an experimental framework.
- Entropix discussion gains momentum: Members engaged in discussions about Entropix, appreciating an overview that offers insight into its operation and potential.
- As the name and concept gain traction, users express anticipation for forthcoming evaluation features.
- Exploring Retrieval-Augmented Generation (RAG): Discussion focused on RAG Techniques, highlighting a GitHub repository that showcases various advanced methods for integrating retrieval with generative models.
- Participants are keen on optimizing performance and understanding the advantages of frameworks like Haystack versus implementing solutions from scratch.
- Jensen's insights on NVIDIA resonate: In an interview, Jensen emphasized NVIDIA's full stack approach and the complex demands of AI infrastructure, referring to the need for accelerated computing across various workloads.
- He acknowledged the transformative impact of generative AI while outlining NVIDIA's competitive position and the necessity for ongoing innovation.
- Challenges of implementing frameworks: Some users expressed skepticism regarding frameworks like Haystack, preferring custom solutions but acknowledged the benefits for less experienced developers.
- They agreed that every framework has its pros and cons, and there is notably critical sentiment surrounding Langchain.
Links mentioned:- Diffusion for World Modeling: Visual Details Matter in Atari (DIAMOND) 💎: Diffusion for World Modeling: Visual Details Matter in Atari (DIAMOND) 💎 Webpage
- Tweet from Dario Amodei (@DarioAmodei): Machines of Loving Grace: my essay on how AI could transform the world for the better https://darioamodei.com/machines-of-loving-grace
- OpenAI Presents: no description found
- Tweet from Ilan Bigio (@ilanbigio): sharing all the meta-prompts and meta-schemas we used for this feature... just published them in the @openai docs check them out, make them your own – would love to hear your questions and feedback! ...
- 1.5M views · 20K reactions | GTA 5 Real Life Graphics | Generative AI | GTA 5 Real Life Graphics | Generative AI | By TRCK | Facebook: GTA 5 Real Life Graphics | Generative AI
- Tweet from Latent.Space (@latentspacepod): 🆕 Production AI Engineering starts with Evals https://latent.space/p/braintrust A 2 hour deep dive on the state of the LLM Ops industry with @ankrgyl following the @braintrustdata Series A! We di...
- Tweet from Yuhuai (Tony) Wu (@Yuhu_ai_): Three components of Reasoning for AI: 1. Foundation (Pre-training) 2. Self-improvement (RL) 3. Test-time compute (planning). @xai will soon have the best foundation in the world - Grok3. Join us to...
- Tweet from attunewise.ai (@attunewise): By "contender" meaning for the next token. From my observations so far, high varentropy is indicative of hallucination (which is what you'd expect) @_xjdr
- AI Legal Tech Startup EvenUp Raises $135M To Hit Unicorn Status: EvenUp, a legal tech startup creating artificial intelligence products for the personal injury sector, raised a $135 million Series D at a valuation of more than $1 billion in the round led by Bain Ca...
- Long-Context LLMs Meet RAG: Overcoming Challenges for Long Inputs in RAG: Retrieval-augmented generation (RAG) empowers large language models (LLMs) to utilize external knowledge sources. The increasing capacity of LLMs to process longer input sequences opens up avenues for...
- Notion – The all-in-one workspace for your notes, tasks, wikis, and databases.: A new tool that blends your everyday work apps into one. It's the all-in-one workspace for you and your team
- Cerebras Inference: no description found
- Tweet from shyamal (@shyamalanadkat): introducing swarm: an experimental framework for building, orchestrating, and deploying multi-agent systems. 🐝 https://github.com/openai/swarm
- Tweet from Clark Tang (@_clarktang): One of the coolest things about Altimeter is how open @altcap encourages us to be with our research process. In this interview w/ Jensen we asked him the questions we (as analysts) wanted to know - a...
- Tweet from swyx @ NYC (@swyx): [Livestream] OpenAI Swarm + Realtime API first looks https://x.com/i/broadcasts/1RDGlyoybRRJL
- Tweet from Tom Dörr (@tom_doerr): OpenAI is thinking about adding DSPy support
- Tweet from Philipp Schmid (@_philschmid): This came unexpected! @OpenAI released Swarm, a lightweight library for building multi-agent systems. Swarm provides a stateless abstraction to manage interactions and handoffs between multiple agents...
- Tweet from Jimmy Apples 🍎/acc (@apples_jimmy): It is time
- Tweet from shyamal (@shyamalanadkat): ‼️ since this started trending unexpectedly: swarm is not an official openai product. think of it more like a cookbook. it’s experimental code for building simple agents. it's not meant for produc...
- Tweet from Adam.GPT (@TheRealAdamG): What I love most about Swarm is its simplicity. It doesn’t try to do too much, but rather just enough. It’s our example to help you get started. Check out the cookbook as the place to get started:...
- GSM-Symbolic: Understanding the Limitations of Mathematical Reasoning in Large Language Models: Recent advancements in Large Language Models (LLMs) have sparked interest in their formal reasoning capabilities, particularly in mathematics. The GSM8K benchmark is widely used to assess the mathemat...
- Tweet from Santiago (@svpino): Large Language Models don't reason. Thank you, Apple.
- GitHub - NirDiamant/RAG_Techniques: This repository showcases various advanced techniques for Retrieval-Augmented Generation (RAG) systems. RAG systems combine information retrieval with generative models to provide accurate and contextually rich responses.: This repository showcases various advanced techniques for Retrieval-Augmented Generation (RAG) systems. RAG systems combine information retrieval with generative models to provide accurate and cont...
Latent Space ▷ #ai-announcements (1 messages):
Production AI Engineering
LLM Ops industry
Evals centrality
Impossible Triangle of LLM Infra
Open source stats
- Production AI Engineering drops new insights: The latest podcast episode features a deep dive on Production AI Engineering with insights from @ankrgyl after the @braintrustdata Series A.
- Discussion includes why Evals are essential in production AI, marking a significant highlight in the industry landscape.
- Evals at the core of LLM Ops: Experts agree that Evals are central to the map of production AI engineering, emphasizing their importance in LLM operations.
- This insight is reinforced by @HamelHusain's contributions, noting the growing focus on evaluation metrics.
- Exploring the Impossible Triangle of LLM Infra: Swyx shared his pet theory on the Impossible Triangle of LLM infrastructure, presenting new perspectives on its implications.
- This led to an updated version of @doppenhe's AI SDLC chart, providing clarity on ongoing discussions within the community.
- Staggering open source stats revealed: The podcast teased shocking statistics revealing that less than 5% of the current LLM models are open source.
- Further details are expected in upcoming discussions, as this figure highlights the trend toward proprietary solutions in AI.
Link mentioned: Tweet from Latent.Space (@latentspacepod): 🆕 Production AI Engineering starts with Evals https://latent.space/p/braintrust A 2 hour deep dive on the state of the LLM Ops industry with @ankrgyl following the @braintrustdata Series A! We di...
Latent Space ▷ #ai-in-action-club (65 messages🔥🔥):
Live Coding Demos
Programming Languages Comparison
Presentation Techniques
DSLs in Programming
Upcoming Sessions
- Live Coding Demos Spark Enthusiasm: Members expressed interest in live coding demos, with one offering to present about building DSLs using Lua, Golang, or Ruby next week.
- Live demos are challenging, and there's a consensus on the need for preparation and planning ahead of time.
- Discussion on Programming Languages: There was a discussion about the effectiveness of various programming languages for coding tasks, with Python being labeled as awful for DSLs.
- Members noted that Lua, Golang, and Ruby could be more suitable choices, emphasizing the unique strengths and downsides of each.
- Presentation Preparation Ideas: Several members showed interest in creating and sharing presentations, with specific mention of building presentations using O1.
- One member offered to present on techniques for conducting live coding presentations, highlighting the need for effective communication during such events.
- Collaborative Workshop Approach: Members agreed on the idea of working together in the chat to create and refine presentations over the coming week.
- This 'pair coding without the pairs' concept aims to foster feedback and collaboration among peers for better presentation outcomes.
- Upcoming Session Announcements: Links for upcoming sessions were shared, including topics such as UI/UX patterns for GenAI and RAG Architectures for effective information retrieval.
- Resources like articles and videos were included to provide further context and material for participants.
Link mentioned: AI In Action: Weekly Jam Sessions: no description found
Stability.ai (Stable Diffusion) ▷ #general-chat (119 messages🔥🔥):
Stable Diffusion Compatibility with 3060ti
Embedding vs Lora Training
Upscaling Techniques
Image to 3D Model Generation
Product Photo Integration
- 3060ti Good for Stable Diffusion: Users discussed the performance of the 3060ti for Stable Diffusion, with some suggesting it performs well despite having only 8GB of VRAM.
- One user shared that using Flux image generation demonstrated the card's capabilities despite its limitations.
- Embedding vs Lora Training: A discussion emerged on the differences between training an embedding and a Lora, with users stating that Lora training typically yields better quality images.
- It was noted that Lora training allows for more detailed diffusion model training compared to embedding which only affects the text encoder.
- Upscaling Techniques Compared: The conversation evaluated the merits of Tiled Diffusion versus Ultimate SD Upscale for image enhancement, highlighting that they serve different purposes.
- While Tiled Diffusion focuses on reducing VRAM consumption, the Ultimate SD Upscale is dedicated solely to improving image resolution.
- Challenges in Image to 3D Model Generation: Participants exchanged thoughts on the complexities of image to 3D model generation, stating that effective solutions are still lacking.
- Users highlighted that techniques like multi-view inference are currently the best available methods to overcome certain challenges in this area.
- Seeking Help with Product Photo Integration: A user requested assistance with integrating product photos into new backgrounds while wanting high-quality results rather than basic compositing.
- Suggestions included the need for specialized training techniques like Lora to produce better-blended images.
Links mentioned:- Webui Installation Guides: Stable Diffusion Knowledge Base (Setups, Basics, Guides and more) - CS1o/Stable-Diffusion-Info
- FLUX 1 NF4 basic workflow + img2img for lowend GPU for ComfyUI: dᴉlɟ ǝɥʇ uᴉddᴉlɟ oqɹnʇ
LlamaIndex ▷ #announcements (1 messages):
LlamaIndex Hackathon
Slides and Resources
- LlamaIndex Hackathon Announcement: A member shared details about the upcoming LlamaIndex Hackathon happening this weekend.
- They provided a link to access all relevant slides and resources for participants.
- Resources for Participants: Participants are encouraged to check out the provided resources to prepare for the hackathon.
- The slides and resources include important information to support their projects.
Link mentioned: Notion – The all-in-one workspace for your notes, tasks, wikis, and databases.: A new tool that blends your everyday work apps into one. It's the all-in-one workspace for you and your team
LlamaIndex ▷ #blog (4 messages):
RAG pipeline with LlamaIndex
Advanced RAG deployment
Multi-agent workflow for business analysis
Multi-agent system for RFP response generation
- Build a RAG Pipeline with LlamaIndex: Check out this video from @awscloud Developers on building a RAG pipeline with LlamaIndex that covers basic workflows and components.
- It includes a simple RAG implementation and the router query technique for enhanced accuracy.
- Simple 3-Step Advanced RAG Process: Deploying advanced RAG can be cumbersome, but it can be simplified to a 3-step process: write your workflow in Python, deploy with llama_deploy, and run it!
- An excellent tutorial by @pavan_mantha1 guides you through the steps.
- Multi-Agent Workflow for Business Analysis: Discover a tutorial on building a multi-strategy business analysis workflow that includes company history, market analysis, and strategy canvas.
- This comprehensive approach is shared in a thread by Lakshmi worth checking out.
- Agentic Workflow for RFP Response Generation: Explore a brand new guide on how to develop a multi-agent system for generating responses to RFPs based on your knowledge base.
- This agentic workflow can seamlessly produce complete responses, as detailed in this release.
LlamaIndex ▷ #general (65 messages🔥🔥):
DocumentSummaryIndex Filters
RouterQueryEngine with Chat History
Langfuse Integration Issues
Image Extraction from PDFs
Colpali Implementation in LlamaIndex
- Filters not applicable to DocumentSummaryIndex: A user expressed difficulty in applying filters to the
DocumentSummaryIndex
, noting that the retriever doesn't pass kwargs correctly.- Another member suggested using a different approach, indicating it might not be necessary to use the Document Summary Index for their use case.
- Integrating Chat History with RouterQueryEngine: A user inquired about the feasibility of using
RouterQueryEngine
with chat history by including all chat messages.- Suggestions were made to utilize workflows for better integration, with links to examples provided.
- Langfuse Integration Troubleshooting: A user reported problems with data visibility in Langfuse while integrating with LlamaIndex, suspecting it might be due to their region.
- Another member recommended ensuring that the code to flush events to Langfuse was included in their application to resolve the issue.
- Challenges in Extracting Images from PDFs: A user raised concerns about extracting images from PDFs, mentioning unexpected ASCII characters in the output.
- They sought guidance on this matter and were looking for clarity regarding the export of parsed data.
- Exploring Colpali Implementation in LlamaIndex: A user asked about the possibility of implementing Colpali within LlamaIndex, noting a lack of documentation.
- Responses indicated that while a full embedding model for Colpali is not currently supported, there is interest in adding it as a reranker in the future.
Links mentioned:- Open Source Observability for LlamaIndex - Langfuse: Open source observability for LlamaIndex. Automatically capture detailed traces and metrics for every request of your RAG application.
- Workflows - LlamaIndex: no description found
- Observability - LlamaIndex: no description found
- Langfuse Callback Handler - LlamaIndex: no description found
- llama_index/llama-index-core/llama_index/core/chat_engine/context.py at main · run-llama/llama_index: LlamaIndex is a data framework for your LLM applications - run-llama/llama_index
- Router Query Engine - LlamaIndex: no description found
- OpenAI - LlamaIndex: no description found
- Cookbook LlamaIndex Integration (Instrumentation Module) - Langfuse: Example cookbook for the experimental LlamaIndex Langfuse integration using the instrumentation module of LlamaIndex.
- Issues · run-llama/llama_index: LlamaIndex is a data framework for your LLM applications - Issues · run-llama/llama_index
tinygrad (George Hotz) ▷ #general (45 messages🔥):
Add Type Annotations in Tinygrad
Bounties and Contribution Requirements
SHA256 Implementation in Tinygrad
MLPerf Inference and StableDiffusion
Meeting Agenda and Updates
- Evaluating Add Type PRs: The discussion centered around evaluating three PRs for adding type annotations, with one rule out due to performance concerns and another by a known contributor being weighed more heavily.
- A final PR was dismissed due to failing tests, raising concerns about unnecessary changes and the overall merging status of the PRs.
- Contribution Requirements for Bounties: George noted that unless contributors have had several PRs merged, they are unlikely to solve a bounty, highlighting a clear progression to becoming a recognized contributor.
- A $200 bounty was added for parallel SHA3 implementation, emphasizing the importance of experience before tackling larger tasks.
- SHA256 Implementation Challenges: A member proposed a complete implementation of SHA256 in Tinygrad while discussing the possibility of parallel processing, although concerns about the current design's limitations were raised.
- George expressed a desire for parallel capabilities in the implementation, reflecting on performance aspects.
- Inquiry about MLPerf: A newcomer inquired about the significance of MLPerf in relation to existing examples in the tinygrad repository, leading to a suggestion to reference existing documentation.
- This points to focus areas within Tinygrad related to optimization and benchmarking.
- Upcoming Meeting and Agenda Items: An upcoming meeting is set to cover various updates from Tiny Corp, performance results, and active bounties, illustrating ongoing development efforts.
- Topics include cloud strategies and driver debugging, indicating a collaborative effort to address potential challenges.
Links mentioned:- Add type annotations to classes in `nn/__init__.py` by mszep · Pull Request #7005 · tinygrad/tinygrad: With this change, mypy --strict tinygrad/nn/__init__.py only complains about return values of functions in other modules.
- refactor _make_hip_code_for_op into pm rules by ignaciosica · Pull Request #7001 · tinygrad/tinygrad: process_replay has diff as casting is no longer fixed and is now a uop
- nn/__init__.py type annotations by Ry4nW · Pull Request #7003 · tinygrad/tinygrad: no description found
- add types in all nn/init.py classes by bhavyagada · Pull Request #7002 · tinygrad/tinygrad: Adds type annotations to all classes and functions in the nn/init.py file. Additionally, type annotations have been included for variables that do not require new lines.
tinygrad (George Hotz) ▷ #learn-tinygrad (11 messages🔥):
Action Diffusion in Tinygrad
DDPM/DDIM Scheduling Library
Gradient Axis Matching in Tensors
Lean Proof Development
Robotic Control with Action Diffusion
- Action Diffusion in Tinygrad needs better docs: A user highlighted issues with the Action Diffusion library in Tinygrad, noting code that mandates a translation back to numpy and adds noise incorrectly when wrapped in
TinyJit
.- They plan to create a more compatible diffuser scheduler next week for better integration with Tinygrad projects.
- User creates DDPM scheduler for Metal training: A member announced they developed their own DDPM scheduler as they could not find an existing library in Tinygrad, making it available for training diffusion models on Metal.
- They are open to sharing their work with others in the community who need it.
- Tensor gradient axis mismatch problem: Discussion arose around strategies for resolving a gradient axis mismatch when updating a tensor 'w' and its gradient, where the user proposed multiple potential solutions.
- Options included aligning axis, removing constraints, or resharding, though the latter was deemed wasteful.
- Challenges with Lean Proof in merging development: A member shared frustrations with developing a Lean proof for merging, citing difficulties with Lean4 syntax and lack of resources compared to Lean3.
- They invited others to team up to make the learning process of Lean4 easier, offering to share their logical frameworks.
- Exploring robotics through action diffusion: A user expressed enthusiasm about combining robotic control with action diffusion, specifically mentioning intentions to explore concepts like Vq-BET and ACT + ALOHA.
- They also extended an invitation for collaboration on getting diffusers operational, especially with ResNet18 backgrounds.
Link mentioned: GitHub - mdaiter/diffusion_action_policy: Action Diffusion, in Tinygrad: Action Diffusion, in Tinygrad. Contribute to mdaiter/diffusion_action_policy development by creating an account on GitHub.
Interconnects (Nathan Lambert) ▷ #news (4 messages):
Replicating OpenAI's O1 Model
Journey Learning Paradigm
Dowehaveopeno1.com Initiative
- Insights from Replicating OpenAI's O1 Model: The first in-depth report on replicating OpenAI's O1 highlights a new training paradigm called 'journey learning', which integrates search and learning in mathematical reasoning with just 327 training samples.
- Zhen Huang noted that this approach includes trial-and-error and backtracking, leading to an 8% improvement in performance.
- Documentation of the Replication Journey: The report meticulously documents the team's discoveries, challenges, and innovative methods encountered during the replication journey.
- Discussion on Dowehaveopeno1.com: A suggestion was made to create the site dowehaveopeno1.com, showing interest in consolidating information related to OpenAI's models.
- However, another member remarked that while it's progress, it's not time to implement the idea just yet.
- Concerns about Twitter Policing: One member expressed a sentiment of not being active enough on Twitter to monitor discussions effectively.
- This points to a wider concern regarding managing community discussions and information dissemination.
Link mentioned: Tweet from Pengfei Liu (@stefan_fee): The first in-depth technical report on Replicating OpenAI's o1 !!! Uncover a Treasure Trove of Trial-and-Error Insights and Hard-Won Lessons. Some highlights: (1) We introduce a new training par...
Interconnects (Nathan Lambert) ▷ #ml-drama (25 messages🔥):
.io domain implications
Microsoft AI researcher joins OpenAI
Benchmark performance of o1-turbo-mini
AGI definitions and discussions
AI researchers' financial success
- .io Domain Transition Sparks Discussion: The geopolitical impact of the British government transferring sovereignty over an island has led to speculation about the future of the .io domain suffix, prompting discussions about Big Tech potentially intervening.
- An article linked highlights how historical events can disrupt digital spaces, paralleling the current situation with past geopolitical shifts.
- Microsoft's Top AI Researcher Moves to OpenAI: News broke that Sebastien Bubeck, one of Microsoft’s leading AI researchers, is moving to OpenAI as reported by various sources, including The Information.
- The move sparked humor and skepticism about the motivations behind such career changes, especially given the lucrative nature of AI roles today.
- o1-turbo-mini Impresses with Benchmarks: There's chatter around the o1-turbo-mini model performing exceptionally well on benchmarks, which some find unexpectedly impressive.
- Comments also reflect that such performance can provoke playful reactions toward those skeptical of AI advancements.
- Debate on AGI and Its Implications: A member critically noted that Bubeck's
- The conversation shifted towards the ambiguous definitions of AGI, with agreements on the nuances involved in the discussions.
- AI Researchers Cashing In: Amid the AI boom, there's a noticeable trend of top researchers making significant financial gains, leading to some playful mockery from those with less financial success in the field.
- This contrast in fortunes has fueled banter, as industry changes rapidly create wealth for a select few.
Links mentioned:- The Disappearance of an Internet Domain: How geopolitics can alter digital infrastructure
- Tweet from Amir Efrati (@amir): news: one of Microsoft's top AI researchers, @SebastienBubeck, is moving to OpenAI. https://www.theinformation.com/briefings/microsoft-ai-researcher-sebastien-bubeck-to-join-openai?rc=c48ukx
Interconnects (Nathan Lambert) ▷ #memes (6 messages):
OpenAI crawling restrictions
Dario Amodei's writings
OpenAI spinoffs
Rise of startups by ex-OpenAI employees
Corporate dynamics in AI
- OpenAI can't crawl Dario Amodei's site: A notable comment surfaced that the only bot not allowed to crawl Dario Amodei's website is OpenAI, which raised some eyebrows.
- This has led to humorous speculation about transparency and accessibility within AI research.
- Dario Amodei's piece impresses: A member praised Dario's writing, stating that his piece is brilliantly written and speculated that it would perform well on lmsys due to its markdown use.
- Such high regard for the content suggests a growing interest in quality academic discussions in AI.
- New OpenAI spinoff rumors: Speculation about yet another OpenAI spinoff has emerged, noted for being beneficial for consumers.
- This points to a potential trend of growth in the AI startup ecosystem stemming from major players' alumni.
- Ex-OpenAI employees launching startups: It's reported that there will soon be 1,700 startups founded by former OpenAI employees, marking significant industry activity.
- This surge in entrepreneurial ventures hints at both innovation and the diversification of the AI landscape.
- OpenAI's impact on corporate structures: There’s a bold claim that OpenAI is not only innovating AI but also causing leftist fracturing in the American corporate world.
- This commentary reflects deeper societal implications of AI advancements and corporate governance.
Links mentioned:- Tweet from Luca Soldaini 🎀 (@soldni): the only bot not allowed to crawl @DarioAmodei's website is @OpenAI 💀
- Tweet from Sheel Mohnot (@pitdesi): Another OpenAI spinoff coming? Great for consumers
- Tweet from Aidan McLau (@aidan_mclau): dario's piece is brilliantly written. it uses a lot of markdown tho. it feels like it'd perform well on lmsys. wait...
Interconnects (Nathan Lambert) ▷ #reads (8 messages🔥):
Machines of Loving Grace
Positive AI Future
Opus Release Hints
- Machines of Loving Grace Gets Rave Reviews: The essay titled Machines of Loving Grace has been described as having a perfect title, capturing attention immediately.
- One member remarked that the beginning is awesome and engaging, drawing interest in the discussion on the potential of AI.
- Industry Leaders Highlight AI's Upside: A member expressed relief that industry leaders are finally discussing the upside of AI, stating they are tired of the fearmongering surrounding the topic.
- This sentiment signals a growing interest in positive conversations about AI's future potential.
- Hints of Opus on the Horizon: A user hinted that Opus may be coming soon, indicating possible upcoming developments in AI.
- This speculation adds excitement to the ongoing discussions about future AI advancements.
Link mentioned: Dario Amodei — Machines of Loving Grace: How AI Could Transform the World for the Better
Interconnects (Nathan Lambert) ▷ #retort-podcast (8 messages🔥):
Vijay Pande and the Nobel Chemistry
Impact of Folding@Home
EVfold and Protein Folding Algorithms
Docking in Drug Discovery
Historical Neural Networks Paper
- Vijay Pande's Surprising Absence from Nobel Chemistry: A member expressed surprise that Vijay Pande didn't receive a Nobel in Chemistry, noting that many AI colleagues at Schrödinger are part of his group.
- One noted the potential impact of his work on CUDA at NVIDIA and questioned how much it has influenced biology through initiatives like Folding@Home.
- Folding@Home: An Early Yet Underwhelming Impact: A member suggested that Folding@Home may not have had the impact it deserved, arguing it was too early to fully realize its potential.
- They highlighted DeepChem as another significant contribution by Pande, driving data science in quantum chemistry.
- EVfold: A Missing Acknowledgment in Protein Folding: There was a discussion about AlphaFold and other protein folding algorithms relying on co-variation of amino acids, initially introduced by EVfold in 2011.
- One member lamented that the Nobel seemed to favor final products over foundational scientific ideas.
- Dominance of Docking Ignored in Nobel Discussion: Another member noted that the significant role of docking in lead drug discovery received inadequate recognition during the Nobel discussions.
- They highlighted that this concept has been extensively used, yet its importance was not acknowledged.
- Historical Paper on Neural Networks from 1982: A member referenced a paper on neural networks by John Hopfield published in PNAS back in 1982, sharing a link to it.
- A comment followed, referring to it as an amusing citation from the 'boomer' era.
DSPy ▷ #show-and-tell (5 messages):
Next.JS voice interview prep platform
FastAPI route for signatures
CLI command generation using Typer
GeneratePureJSX signature
GenerateMarkdown signature
- Next.JS Voice Interview Prep Platform Launch: A member announced the development of a full stack Next.JS voice interview prep/testing platform in the lounge.
- This innovative platform aims to enhance interview preparation through voice interaction.
- FastAPI Route Creation for Signatures: A member shared a code snippet that turns any dspy.Signature into a FastAPI route, allowing it to return the predictor as a dictionary.
- The implementation utilizes the init_instant function to initialize the environment for processing requests.
- Dynamic CLI Command Creation with Typer: A code example converting dspy.Signature classes into CLI commands using Typer was shared, demonstrating automation.
- The feature dynamically builds the command function based on input fields specified in the signature class.
- GeneratePureJSX Signature Demonstrated: The GeneratePureJSX signature is showcased, designed to generate clean JSX code for React environments.
- This signature includes fields for context and requirements, producing well-structured output.
- GenerateMarkdown Signature Highlighted: Another signature, GenerateMarkdown, was introduced to create markdown text based on a given title and content.
- Similar to GeneratePureJSX, this provides a defined structure with input fields for generating markdown.
Links mentioned:- Introduction to DSLModel Framework: https://github.com/seanchatmangpt/dslmodel Hi everybody, Sean Chatman here, introducing the DSL Model Framework, a powerful tool for modeling data using DSPy and Jinja. In this video, I explain how t...
- - YouTube: no description found
- GitHub - seanchatmangpt/dslmodel: Structured outputs from DSPy and Jinja2: Structured outputs from DSPy and Jinja2. Contribute to seanchatmangpt/dslmodel development by creating an account on GitHub.
DSPy ▷ #papers (4 messages):
GraphIC for ICL selection
Inference-time computation impacts on LLMs
StructRAG in knowledge-intensive tasks
- GraphIC transforms ICL selection: The paper introduces GraphIC, a technique using graph-based representations and Bayesian Networks (BNs) to enhance the selection of in-context examples (ICEs) for large language models (LLMs). It addresses the limitations of conventional text-based embedding methods in multi-step reasoning tasks.
- Graph structures filter out shallow semantics, preserving deeper reasoning structures necessary for effective multi-step reasoning.
- Enhancing LLMs via inference-time computation: Research on inference-time computation investigates how allowing LLMs to utilize fixed additional compute can boost performance on difficult prompts. The findings highlight important implications for future LLM pretraining and the balance between inference-time and pre-training compute.
- Current studies reveal that understanding the scaling behaviors of test-time inference methods is critical, though many strategies have shown negative results.
- StructRAG optimizes retrieval in reasoning tasks: The paper proposes a new framework called StructRAG to enhance retrieval-augmented generation (RAG) methods, focusing on better handling of knowledge-intensive reasoning tasks. By converting raw information into structured knowledge, the framework aims to effectively identify needed information.
- This structured approach can optimize information for various tasks, facilitating improved reasoning amidst the noisy augmentation traditionally faced in existing RAG methods.
Links mentioned:- StructRAG: Boosting Knowledge Intensive Reasoning of LLMs via Inference-time Hybrid Information Structurization: Retrieval-augmented generation (RAG) is a key means to effectively enhance large language models (LLMs) in many knowledge-based tasks. However, existing RAG methods struggle with knowledge-intensive r...
- GraphIC: A Graph-Based In-Context Example Retrieval Model for Multi-Step Reasoning: In-context learning (ICL) enables large language models (LLMs) to generalize to new tasks by incorporating a few in-context examples (ICEs) directly in the input, without updating parameters. However,...
- Scaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model Parameters: Enabling LLMs to improve their outputs by using more test-time computation is a critical step towards building generally self-improving agents that can operate on open-ended natural language. In this ...
DSPy ▷ #general (18 messages🔥):
LLM Classifier Training
Versioning Documentation
Creative Writing Criteria Optimization
Contextual Embedding in DSPy
Metric Programs for Evaluation
- LLM Classifier Seeks Ambiguity Handling: A user is working on training an LLM classifier and is seeking community input on handling classification ambiguities, suggesting a method for the model to output when it encounters uncertainty.
- Should I create separate classes for all the possible ambiguities? was the question raised.
- Community Suggests to Enhance LLM Classifier: The response to the ambiguity issue advised against creating new classes, instead recommending adding a second output field in the LLM signature to declare ambiguities.
- One member even offered to prototype a small example to assist in this implementation.
- Versioning Docs for Better Clarity: A new user requested versioned documentation in DSPy, stating it would help clarify usage regarding changes reflected in docs that aren't yet released.
- The community member acknowledged this as a wise idea and noted that the latest release is out now.
- Exploration of Explicit Criteria in Optimizers: There's a discussion about developing an optimizer that could identify explicit criteria for tasks to illuminate patterns in outputs based on human labels.
- A detailed approach was suggested involving a program that uses pattern matching to generate criteria relevant to success or failure in creative writing.
- Need for Guidance on Contextual Embedding: A user inquired about a cookbook or guide on using contextual embeddings within a DSPy program, referencing an external resource.
- The user expressed confusion and requested assistance to better understand the implementation.
Links mentioned:- Tweet from Omar Khattab (@lateinteraction): when chatgpt manages to finally crack a good joke:
- dspy-examples/src/summarization/programs/metric at main · Scale3-Labs/dspy-examples: A collection of example AI programs built using DSPy and maitained by the Langtrace AI team. - Scale3-Labs/dspy-examples
DSPy ▷ #examples (9 messages🔥):
Chatbot Features Extraction
Off-topic Classification
Input Training Set
Cosine Similarity Metric
- Stuart's chatbot feature extraction process: Stuart is working on building a chatbot that extracts features, including identifying off-topic queries through a database of criteria.
- This approach includes an off-topic table with entries like Insufficient Info, which flags queries lacking context.
- Training set for typical inputs: Stuart confirmed having a training set that categorizes queries based on whether they are off-topic or not.
- This dataset aims to improve the response accuracy of the chatbot by filtering irrelevant queries.
- Assessing output effectiveness: okhattab inquired about the development of a metric to evaluate if the chatbot outputs meet the established criteria.
- Stuart mentioned he lacks predefined categories and is considering using cosine similarity to compare input queries with generated categories.
- Exploration for improvement suggestions: Stuart actively seeks suggestions on enhancing his cosine similarity metric approach for assessing outputs.
- His focus is on refining how queries are matched to corresponding categories to improve off-topic detection.
- Community interest in topic discussion: Another member expressed interest in understanding how improvements in off-topic classification could be accomplished.
- This indicates a broader curiosity within the community on the methods of enhancing chatbot effectiveness.
LAION ▷ #general (18 messages🔥):
LLaMA 3.2
PixTral Comparison
Epic Games & Sketchfab
Advanced Cookie Monster Mode
CogVLM Output
- LLaMA 3.2 excels in pop culture knowledge: Discussion highlighted that LLaMA 3.2 outperforms competitors in pop culture knowledge due to its training on 5 billion images, making its captions more coherent.
- Compared to models like Molmo or PixTral, LLaMA 3.2 provides better contextual understanding in its imagery captions.
- PixTral best for adult content: Members noted the unique positioning of PixTral, primarily beneficial when trained on adult content, while LLaMA 3.2 excels in general contexts.
- It's emphasized that LLaMA 3.2 is preferred in side-by-side comparisons for broader cultural relevance.
- Impact of Epic Games removing Sketchfab: Epic Games' decision to remove Sketchfab will lead to the loss of 800k 3D models from the Objaverse, urging members to download them quickly.
- The announcement sparked conversations about the potential implications for 3D modeling communities and users who rely on that resource.
- Advanced Cookie Monster Mode discussed: A YouTube link titled 'Advanced Cookie Monster Mode :D' was shared, promoting its intriguing content but lacking further details.
- No additional commentary or context surrounding the video was provided.
- CogVLM shows engagement with cartoon culture: CogVLM was described as capturing the essence of cartoon culture through a vibrant portrayal of characters celebrating the 20th anniversary of Cartoon Network.
- Its output was noted for recognizing a wide range of characters, highlighting the nostalgic value for fans.
Links mentioned:- F5-TTS: no description found
- Advanced Cookie Monster Mode :D: no description found
- CaptionEmporium/laion-pop-llama3.2-11b · Datasets at Hugging Face: no description found
LAION ▷ #research (10 messages🔥):
o1-mini vs o1-preview
Performance of o1 LLM
Diffusion Models as Representation Learners
Representation Alignment
Training T2I Models with CLIP
- o1-mini can't match o1-preview performance: Member shared that o1-mini lacks the capabilities of o1-preview, describing it as brittle on simple problems. They noted that o1-preview outperforms all other SOTA LLMs in robustness, even on Olympiad level tasks here.
- Yet another tale of Rise and Fall highlights claims that o1-mini could handle complex problems but falters on basic reasoning tasks.
- o1 is the best LLM experience: One member emphasized that o1 is the best LLM they've used, effectively handling both papers and code. They claimed that nearly all code generated by o1 works on the first try, showcasing its reliability.
- Insights from REPresentation Alignment paper: A member discussed a paper on generative diffusion models, emphasizing the dramatic improvement in performance through external high-quality representations. They questioned whether the enhancement stemmed from knowledge distillation or a deeper insight regarding generative modeling processes.
- The introduction of REPresentation Alignment indicates notable advancements in training efficiency when connecting hidden layer outputs to neural representations of images.
- Speeding up training with vision encoders: Another member pointed out that aligning outputs of hidden layers with vision encoders speeds up training due to their rich image representations. They noted parallels with training T2I models using CLIP, a text encoder aligned with images.
- Challenges with CLIP's contrastive training: While using CLIP for training T2I models accelerates the process, it also introduces artifacts linked to its contrastive training methods. The member indicated that these artifacts could affect the overall training quality.
Links mentioned:- Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think : Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think
- Tweet from Jenia Jitsev 🏳️🌈 🇺🇦 (@JJitsev): (Yet) another tale of Rise and Fall: o1-mini was claimed to match much larger scale o1-preview on olympiad level math & coding problems. Can it handle simple AIW problems that reveal generaliz...
LLM Agents (Berkeley MOOC) ▷ #mooc-announcements (1 messages):
Graham Neubig's lecture
AI agents in software development
Challenges in real-world software tasks
Multimodal data processing
- Live Lecture Alert: Graham Neubig on AI Agents: Today at 3:00pm PST, our 6th lecture featuring Graham Neubig discussing "Agents for Software Development" will be streamed live here.
- He will cover the state-of-the-art in software development agents, addressing challenges like editing files and multimodal data processing.
- Exploring Challenges in Software Development Tasks: Graham Neubig aims to tackle the complexities of developing AI agents that assist in real-world software tasks, particularly for large repositories.
- Key challenges include identifying which files to edit, testing edits, and integrating web browsing into coding processes.
- Meet the Speaker: Graham Neubig: Graham Neubig is an associate professor at Carnegie Mellon University, specializing in natural language processing research. His focus includes large language models for applications like code generation and question answering.
- Audiences are encouraged to engage with questions in the designated course staff channel for further inquiries.
LLM Agents (Berkeley MOOC) ▷ #mooc-questions (18 messages🔥):
Course Registration Issues
Quiz Google Form Access
Course Timings
Supplemental Course Information
Lab Access Questions
- Course Registration Still Open: New member expressed interest in signing up for the course before the deadline of Dec 12 and was advised that registration is still possible, confirming that the link works for others.
- Another user confirmed their registration was successful after troubleshooting the link.
- Quiz Google Form Troubleshooting: Multiple users reported issues with accessing the Google Forms link for quizzes, but responses indicated that the link worked for others after clearing their cache.
- One participant confirmed they could access the form after troubleshooting, highlighting that it might just be a browser issue.
- Course Timings Revealed: A user inquired about the livestream schedule and was informed that the course timings are from 3:00 PM to 5:00 PM PST.
- This information is crucial for participants needing to plan attendance for live sessions.
- Useful Links and Resources: Course materials and supplemental info, including instructions and additional resources, are confirmed to be available on the course website.
- Members were reminded that they could access links for course sign-up and community chat via the Discord server.
- Lab Access Delays: New participants expressed concerns about not receiving emails regarding access to labs and assignments after submitting the signup form.
- They were directed to the course website for a comprehensive overview of what to expect in terms of lab instructions and assignments.
Link mentioned: Large Language Model Agents: no description found
LLM Agents (Berkeley MOOC) ▷ #mooc-lecture-discussion (7 messages):
Definition of AI Agent
Agentic Applications
Chain of Thought (CoT) Methodology
- Clarifying AI Agent Criteria: An AI agent is characterized by its ability to autonomously execute tasks by interacting with databases or APIs, allowing it to achieve defined goals. For example, while ChatGPT qualifies as an agent, gpt-3.5-turbo remains an LLM model without such capabilities.
- Agentic Applications Explained: An agentic application alters its internal state based on new evidence and acts independently, driven by a reward system. This mechanism allows the app to evolve by reflecting real-world conditions.
- Crowdsourcing AI Agent Definitions: There is an ongoing effort to refine the definition of an AI agent, with contributions and discussions happening on platforms like Twitter. A notable response involved a suggestion to track different versions of the definition collaboratively.
- Understanding CoT's Effectiveness with LLMs: The Chain of Thought (CoT) methodology enhances problem-solving in LLMs by generating intermediate steps that simplify complex problems. By breaking down larger tasks, solutions can be derived from these smaller components.
- Community Discussion on CoT: Members discussed the practical advantages of the CoT approach, emphasizing its ability to facilitate clarity when tackling complex challenges. As underscored by an example with Apple, this strategy effectively aids in arriving at final answers.
Link mentioned: Tweet from Simon Willison (@simonw): Let’s see if we can crowdsource a robust definition of “agent” (with respect to AI and LLMs) that fits in a <=280 character tweet Reply to this with your best attempt, then scroll through the repl...
LLM Agents (Berkeley MOOC) ▷ #mooc-readings-discussion (1 messages):
AI-Powered Search Book
- Essential Read on AI-Powered Search: A member highlighted that this book is expected to be the definitive resource for AI-powered search in the coming years.
- This book will probably be the go-to resource for the next few years.
- Expected Influence of AI-Powered Search Book: The discussion emphasized the potential impact of this book on the field of AI-powered search, forecasting its relevance for future professional practice.
- The member speculated that it would serve as a key reference for AI practitioners and researchers.
Torchtune ▷ #general (10 messages🔥):
Gemma 2 support
Flex Attention implementation
Logit softcapping
CUDA dependencies
Testing considerations
- Gemma 2 Support via Flex Attention: Members discussed implementing Gemma 2 using Flex Attention, noting that the only blocker is logit softcapping which will require an appropriate
score_mod
function to be defined.- It's suggested that a tradeoff for using Flex is acceptable since it simplifies the process, despite potential reliance on CUDA with high compute capabilities.
- Logit Softcapping and Flex APIs: It was noted that while logit softcapping isn't directly available with
F.sdpa
, it can be exposed in_sdpa_or_flex_attention()
through Flex APIs, allowing for its implementation.- The discussion highlighted that shifting to Flex could allow quicker adaptation for upcoming attention operators in future models.
- Testing and Implementation Concerns: Concerns were raised about the testing required for Gemma 2 across different model sizes, with a suggestion to crowdsource testing efforts for those with more computational resources.
- Participants agreed on the importance of ensuring that the implementation is robust even if it means relying on a simpler fallback where needed.
- CUDA Dependency Considerations: There was a consensus that the implementation of Gemma 2 via Flex may be limited by requiring CUDA with sufficient compute capability, highlighting a gap for non-CUDA users.
- Members acknowledged that this restriction would allow for more focused efforts to develop the support features needed incrementally.
- Naive Fallback vs Flex: A member proposed that having a
naive_fallback
could be beneficial to support all functionalities as they emerge, despite the possible overhead in maintenance.- Concerns were shared about whether Flex's flexibility might alleviate these maintenance challenges over time.
Links mentioned:- torchtune/torchtune/utils/_import_guard.py at main · pytorch/torchtune: PyTorch native finetuning library. Contribute to pytorch/torchtune development by creating an account on GitHub.
- pytorch/torch/nn/attention/flex_attention.py at main · pytorch/pytorch: Tensors and Dynamic neural networks in Python with strong GPU acceleration - pytorch/pytorch
Torchtune ▷ #dev (1 messages):
Flex attention on RTX 4090
Shared memory issues in PyTorch
- Facing Out-of-Shared-Memory Problems with Flex Attention: A user shared an issue on GitHub regarding out-of-shared-memory problems while using flex attention on the RTX 4090.
- The issue describes specific errors encountered, suggesting a potential fix for others experiencing similar problems.
- Bug Report on Shared Memory Resources: The GitHub issue outlines a bug where using flex attention results in an out-of-resources error with shared memory.
- It includes a minimal reproduction code snippet and encourages collaboration to resolve the identified problem.
Link mentioned: Shared memory out of resource when using flex attention · Issue #133254 · pytorch/pytorch: 🐛 Describe the bug When I use flex attention on one RTX 4090, I got some error. A minimal repro: import torch from torch.nn.attention.flex_attention import flex_attention flex_attention = torch.com.....
Torchtune ▷ #papers (14 messages🔥):
Aria Model Launch
Pixtral 12B Performance
State of AI Report 2024
LegoScale for Distributed Training
ICLR Submissions on OpenReview
- Aria open multimodal model introduced: The new model, Aria, is an open multimodal native AI that excels in language and coding tasks with a 3.9B and 3.5B parameter architecture.
- It outperforms both Pixtral-12B and Llama3.2-11B, following a comprehensive pre-training process.
- Pixtral 12B faces competition: The Pixtral 12B model is compared to Aria, which shows superior performance but lacks direct benchmark comparisons due to simultaneous releases.
- Members noted that some new benchmarks were better for Aria at launch, highlighting a need for clearer head-to-head comparisons.
- State of AI 2024 Report sparks insights: The State of AI Report 2024 by Nathan Benaich discusses major trends and investments in AI, minimizing mentions of specific models like Torchtune.
- The report serves to inform and inspire discussions on AI's trajectory, particularly impact in fields like medicine and biology.
- LegoScale promises advances in distributed training: LegoScale presents a customizable, PyTorch-native system enabling modular 3D parallel pre-training for large language models, enhancing performance significantly.
- It aims to simplify complex training across multiple libraries and vast GPU ecosystems, potentially revolutionizing distributed training approaches.
- ICLR submissions available on OpenReview: Members discuss how ICLR utilizes OpenReview to access preprints and track reviewer comments throughout the submission process.
- This approach contrasts with NeurIPS, where submissions remain hidden unless publicly shared.
Links mentioned:- State of AI Report 2024: The State of AI Report analyses the most interesting developments in AI. Read and download here.
- Aria: An Open Multimodal Native Mixture-of-Experts Model: Information comes in diverse modalities. Multimodal native AI models are essential to integrate real-world information and deliver comprehensive understanding. While proprietary multimodal native mode...
- LegoScale: One-stop PyTorch native solution for production ready...: The development of large language models (LLMs) has been instrumental in advancing state-of-the-art natural language processing applications. Training LLMs with billions of parameters and trillions...
- Pixtral 12B: We introduce Pixtral-12B, a 12--billion-parameter multimodal language model. Pixtral-12B is trained to understand both natural images and documents, achieving leading performance on various multimodal...
LangChain AI ▷ #general (15 messages🔥):
Community Closure Announcement
Swarm.js Launch
Real Estate LangChain Project
OpenAI API Integration
ImageMessage Type Discussion
- Community Closure Announcement: Jess announced that the LangChain Discord community will close on October 31, 2024, to focus on building a new and improved community. Members are encouraged to fill out a form to stay updated about the changes.
- Feedback is welcomed at community@langchain.dev, and there’s an invitation for anyone interested in becoming a moderator.
- Swarm.js Launch: Pulkitgarg introduced Swarm.js, a Node.js SDK for orchestrating multi-agent systems using OpenAI's API, aimed at the JavaScript community. It includes features for defining agents with custom instructions, making the process simple and ergonomic.
- Developers are encouraged to check out the GitHub repo and contribute to the project.
- Real Estate LangChain Project: Gustaf_81960_10487 shared their open-sourced LangChain real estate project that helped them secure a job, offering to aid contributors with letters of recommendation. They created a GitHub repo for the project and provided a link to their site, Condo Cube.
- Contributors are encouraged to review the repo and get involved in solving real-world issues.
- OpenAI API Integration Inquiry: Mikef0x inquired about plans for integrating OpenAI's real-time API, mentioning issues with Pydantic. Members discussed their experiences with Azure OpenAI and SQLDatabase, noting some humorous outcomes in responses.
- ImageMessage Type Discussion: Grigorij0543 proposed introducing an ImageMessage type in LangChain to simplify image handling in chats. Clay_ferguson commented that text is usually associated with images and suggested the need for a more versatile file association method.
Links mentioned:- Airtable | Everyone's app platform: Airtable is a low-code platform for building collaborative apps. Customize your workflow, collaborate, and achieve ambitious outcomes. Get started for free.
- GitHub - youseai/openai-swarm-node: Swarm.js is a Node.js implementation of OpenAI’s experimental Swarm framework. This SDK allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, while leveraging Node.js for building scalable, real-world applications.: Swarm.js is a Node.js implementation of OpenAI’s experimental Swarm framework. This SDK allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, w...
- GitHub - GGyll/condo_gpt: An intelligent assistant for querying and analyzing real estate condo data in Miami.: An intelligent assistant for querying and analyzing real estate condo data in Miami. - GitHub - GGyll/condo_gpt: An intelligent assistant for querying and analyzing real estate condo data in Miami.
- - YouTube: no description found
LangChain AI ▷ #langserve (1 messages):
Swarm.js
Node.js SDK
OpenAI API
Multi-agent systems
Open-source contributions
- Introducing Swarm.js for Multi-Agent Magic: Announcing Swarm.js, a lightweight Node.js SDK that orchestrates multi-agent systems using the OpenAI API. This framework, inspired by the original OpenAI Swarm in Python, enables easy agent management and task execution.
- With simple orchestration capabilities and customizable agents, developers can seamlessly integrate OpenAI's power into their applications.
- Install Swarm.js Easily: To get started, simply run
npm install openai-swarm-node
to install Swarm.js. This lightweight SDK is designed for practical, real-world use cases.- The system encourages developers of all levels to contribute to the project, promoting community involvement.
- Call for Contributions on GitHub: The project is open for contributions, inviting users to check the GitHub Repo to raise issues and submit pull requests. All developers, whether beginners or experts, are welcome to help enhance the framework.
- Active collaboration and idea sharing are encouraged to build something outstanding together in the open-source community.
Link mentioned: GitHub - youseai/openai-swarm-node: Swarm.js is a Node.js implementation of OpenAI’s experimental Swarm framework. This SDK allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, while leveraging Node.js for building scalable, real-world applications.: Swarm.js is a Node.js implementation of OpenAI’s experimental Swarm framework. This SDK allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, w...
LangChain AI ▷ #langchain-templates (1 messages):
Swarm.js
Multi-agent systems
OpenAI API
Node.js SDK
Contributions to open source
- Swarm.js Launches for Node.js Enthusiasts: Introducing Swarm.js, a lightweight Node.js SDK for orchestrating multi-agent systems using the OpenAI API.
- This framework, inspired by the original Python version, allows developers to easily create and manage agents that perform tasks and communicate with each other.
- Key Features of Swarm.js Unveiled: Swarm.js boasts features like simple orchestration of agents, custom instructions, and a lightweight structure designed for real-world applications.
- Its ergonomic design aims to enhance usability and flexibility for developers looking to harness OpenAI's capabilities.
- Installation and Getting Started: To install Swarm.js, simply run
npm install openai-swarm-node
and begin integrating with the OpenAI API.- The launch encourages both beginners and experts to jump in and explore the framework's potential in their projects.
- Community Contributions Are Welcome!: The project lead invites users to contribute by checking out the GitHub repo, raising issues, and submitting pull requests.
- This collaborative approach aims to enrich the SDK and foster a vibrant community around Swarm.js.
Link mentioned: GitHub - youseai/openai-swarm-node: Swarm.js is a Node.js implementation of OpenAI’s experimental Swarm framework. This SDK allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, while leveraging Node.js for building scalable, real-world applications.: Swarm.js is a Node.js implementation of OpenAI’s experimental Swarm framework. This SDK allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, w...
LangChain AI ▷ #share-your-work (5 messages):
bootstrap-rag v0.0.9 release
Contextual Retrieval using Langchain
Swarm.js Node.js SDK
- bootstrap-rag v0.0.9 is live!: The early release of bootstrap-rag v0.0.9 introduces removal of unwanted test code, along with bug fixes and documentation improvements.
- Key features include LangChain integration, MLflow-evals support, and tested Qdrant templates, enhancing RAG capabilities.
- Exploring Contextual Retrieval: A new YouTube video titled 'Contextual Retrieval using Langchain and OpenAI Swarm Agent' guides viewers through implementing contextual retrieval.
- The video promises insights into using Langchain with OpenAI’s Swarm Agent for effective information retrieval.
- Introducing Swarm.js for Node.js: Swarm.js is launched as a lightweight Node.js SDK for orchestrating multi-agent systems utilizing OpenAI's API, designed for the JavaScript community.
- This framework features simple multi-agent orchestration and is customizable, inviting contributions from developers at all skill levels.
Links mentioned:- Contextual Retrieval using Langchain and OpenAI Swarm Agent: We will take a look at how to implement Contextual Retrieval using Langchain and OpenAI Swarm Agenthttps://github.com/githubpradeep/notebooks/blob/main/conte...
- GitHub - youseai/openai-swarm-node: Swarm.js is a Node.js implementation of OpenAI’s experimental Swarm framework. This SDK allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, while leveraging Node.js for building scalable, real-world applications.: Swarm.js is a Node.js implementation of OpenAI’s experimental Swarm framework. This SDK allows developers to orchestrate multi-agent systems using OpenAI’s API in a lightweight and ergonomic way, w...
- bootstrap-rag: None
- Release v0.0.9 · pavanjava/bootstrap-rag: What's Changed -removed the unwanted testset generation code by @pavanjava in #44 Bug fixes by @pavanjava in #46 Docs enhancement by @pavanjava in #47 Docs enhancement by @pavanjava in #48 Langch...
LangChain AI ▷ #tutorials (1 messages):
LangGraph tutorial
Resume optimization
Interview preparation tools
- LangGraph tutorial launched!: A new tutorial showcases a simple yet powerful two-node LangGraph app that retrieves data from a resume and a job description to answer various questions. Watch the tutorial here.
- The app can re-write resume parts to match job descriptions and generate relevant interview questions.
- Resume and cover letter capabilities: The LangGraph app can also write job-specific cover letters and provide feedback for interviews based on the provided data. This feature makes it easier for job seekers to prepare for applications.
- This tool is particularly helpful for those just starting with LangGraph, as the tutorial is described as easy to follow.
OpenAccess AI Collective (axolotl) ▷ #general (1 messages):
duh_kola: Has anyone tried instruction data during pretraining
OpenAccess AI Collective (axolotl) ▷ #general-help (9 messages🔥):
Config sharing
Sample packing importance
Using adapters for model training
Merging adapters
Training with new data
- Members discuss config sharing: Member @le_mess requested to see the config from @palebluedot_32091, who confirmed its accuracy but mentioned potential improvements.
- At glance it looks correct, said @le_mess, suggesting enabling sample packing in the config.
- Importance of sample packing: Member @le_mess highlighted the need to enable sample packing while running the training command
accelerate launch -m axolotl.cli.train config.yml
.- They also mentioned challenges with multi-GPU setups and using LoRA in earlier discussions.
- Fine-tuning with existing adapter: Member @vortexwault inquired about refining a fine-tuned model using Llama 3 as a base with an adapter for better accuracy.
- @palebluedot_32091 advised merging the adapter into a new base model before training on additional data.
- Guide for merging adapters: When @vortexwault asked for guidance on merging the adapter, @palebluedot_32091 directed them to a GitHub guide.
- He also indicated that the new training run's config should include the line
base_model: '/path/to/merged/model'
.
- He also indicated that the new training run's config should include the line
- Follow-up on guidance: @vortexwault thanked @palebluedot_32091 for the guidance and mentioned he would look into the recommended resource.
- "Let me check it brother thank you," said @vortexwault, indicating his intent to follow the advice.
Links mentioned:- GitHub - axolotl-ai-cloud/axolotl: Go ahead and axolotl questions: Go ahead and axolotl questions. Contribute to axolotl-ai-cloud/axolotl development by creating an account on GitHub.
- GitHub - axolotl-ai-cloud/axolotl: Go ahead and axolotl questions: Go ahead and axolotl questions. Contribute to axolotl-ai-cloud/axolotl development by creating an account on GitHub.
OpenAccess AI Collective (axolotl) ▷ #axolotl-help-bot (5 messages):
Instruct Models
Text Completion Training
Model Specialization
Overfitting Risks
- Instruct models get optimized with text completion training: Training an instruct model, such as GPT-3.5-Instruct, on a text completion task can significantly improve its performance on instruction-based tasks, resulting in more accurate and contextually relevant completions.
- “This can result in improved ability to generate text that aligns closely with prompts.”
- Risk of overfitting in narrow training: There are potential risks of overfitting if the training dataset for the text completion task is not diverse or is too small, leading to high performance on familiar data but poor generalization.
- A caution emphasized was that performance might degrade on other tasks not seen during the specific training.
- Adaptation to new formats and fields: The training can help the model adapt to new formats or types of instructions that it was not exposed to during its initial training.
- This adaptation fosters the model's ability to handle a broader range of inputs more effectively.
- Importance of diverse training datasets: To achieve optimal outcomes when training instruct models, using a diverse and representative dataset is essential.
- Monitoring for overfitting and ensuring broad performance evaluation across tasks is crucial for maintaining generalization capabilities.
Link mentioned: OpenAccess-AI-Collective/axolotl | Phorm AI Code Search: Understand code, faster.
OpenInterpreter ▷ #general (1 messages):
Message Format Compliance
Channel Rules
Links to Documentation
- Message Format Compliance Reminder: Members were reminded to adhere to the prescribed message format in channel <#1210088092782952498>, as per established rules.
- The message emphasized the importance of following guidelines to maintain channel organization and clarity.
- Reference to Channel Rules: There was a reference to the rules outlined in channel <#1207731120138092628>, guiding the expected behavior and contributions in discussions.
- Members are encouraged to review these rules for a better understanding of the channel dynamics.
- Documentation Link Provided: A link was shared directing members to relevant documentation on channel <#1207731120138092628> for further clarity on expected practices.
- The shared link serves as a resource for enhancing compliance and ensuring proper channel communication.
OpenInterpreter ▷ #ai-content (3 messages):
Aria Multimodal Model
User Experiences with Aria
AI Reasoning Limitations
- Aria Takes the Stage as a Game-Changer: The new multimodal model, Aria, by @rhymes_ai_, is now ranked first on the 🤗 Open LLM Leaderboard, featuring 24.9B parameters and capable of handling image, video, and text inputs.
- Users Rave About Aria's Performance: This is the BEST vision language model I have ever tried! claims a user regarding the 25.3B multimodal Aria model, emphasizing its impressive capabilities with image and video inputs.
- The model has been released under the Apache-2.0 license, and fine-tuning scripts are also available to the community.
- AI Reasoning Under Scrutiny: A YouTube video titled 'Apple DROPS AI BOMBSHELL: LLMS CANNOT Reason' explores critical discussions around the reasoning capabilities of language models.
- The creator encourages viewers to prepare for AGI, sharing links to their social media and related content, hinting at a broader conversation about AI's current limitations.
Links mentioned:- Tweet from Alina Lozovskaya (@ailozovskaya): 🌸 The first multimodal model on the 🤗 Open LLM Leaderboard! rhymes-ai/Aria is an open multimodal native model designed to deliver best-in-class performance across vision, language, and coding tasks...
- Apple DROPS AI BOMBSHELL: LLMS CANNOT Reason: Prepare for AGI with me - https://www.skool.com/postagiprepardness 🐤 Follow Me on Twitter https://twitter.com/TheAiGrid🌐 Checkout My website - https://thea...
- Tweet from merve (@mervenoyann): this is the BEST vision language model I have ever tried! Aria is a new model by @rhymes_ai_: a 25.3B multimodal model that can take image/video inputs 🤩 They release the model with Apache-2.0 lice...
AI21 Labs (Jamba) ▷ #general-chat (2 messages):
Jamba support thread
- Support Inquiry on Jamba Issues: A member created a thread in <#1222916247063232553> regarding an issue experienced while trying to run Jamba and asked if that was the correct way to seek support.
- Keepitirie responded, confirming they addressed the member's query in that channel and encouraged continued discussion there.
- Continuing the Discussion in Channel: Another member suggested that the discussion about the Jamba issue should remain in the original thread for clarity and coherence.
- They emphasized the importance of following up in the same channel to ensure all relevant information is easily accessible.
Mozilla AI ▷ #announcements (1 messages):
Community Building
Panel Discussion on Engagement
Expert Insights
Sustainable Community Models
- Panel Discussion Unveiled for Community Building: A distinguished panel of Community and Developer Relations experts is set to discuss actionable strategies for fostering community engagement and project success.
- Attendees can gain insights from experts including Jillian Bejtlich and Rynn Mancuso, focusing on techniques to grow user bases and increase contributions.
- Strengthen Your Community Skills: The panel will provide tactical advice on how to build a thriving community around projects, emphasizing relationship-building beyond just writing code.
- This is an invaluable opportunity for project leads to enhance their community-building skills and connect with others in the field.
- Don't Miss Out on this Event!: Everyone is encouraged to RSVP for this insightful panel discussion on community-building practices here.
- “Don’t miss out on this invaluable opportunity!” echoes through the channel encouraging active participation.
DiscoResearch ▷ #general (1 messages):
mihai4256: Starting from the above, I made this: https://github.com/Mihaiii/backtrack_sampler
Gorilla LLM (Berkeley Function Calling) ▷ #discussion (1 messages):
Multi-turn Examples Evaluation
Model Performance Metrics
Modification Restrictions on Base Handler
- Challenges in Multi-Turn Evaluation Process: Members discussed the evaluation method for multi-turn examples, trying each turn 20 times before concluding an attempt as a fail if the model doesn't output an empty string.
- A significant issue arises since even with correct predictions, the model cannot break the count loop resulting in a ~0% performance.
- Temporary Workaround Improves Performance: One member noted that modifying the code in base_handler.py to only try once per round leads to an improvement to ~15% performance in multi-turn evaluations.
- However, they emphasized that modifications to base_handler.py are not permitted, seeking alternative resolutions.
- Tweet from Phi Hoang (@apostraphi)uts. - It boasts a **64k token context window** and was pre-trained on a diverse dataset, including **400B multimodal tokens**, ensuring stable performance across tasks ([Paper](https://hf.co/papers/2410.05993), [Blog](https://rhymes.ai/blog-details/aria-first-open-multimodal-native-moe-model)). - **Users Rave About Aria's Performance**: *This is the BEST vision language model I have ever tried!* claims a user regarding the **25.3B multimodal Aria model**, emphasizing its impressive capabilities with image and video inputs. - The model has been released under the **Apache-2.0 license**, and fine-tuning scripts are also available to the community. - **AI Reasoning Under Scrutiny**: A [YouTube video](https://www.youtube.com/watch?v=tTG_a0KPJAc) titled 'Apple DROPS AI BOMBSHELL: LLMS CANNOT Reason' explores critical discussions around the reasoning capabilities of language models. - The creator encourages viewers to prepare for AGI, sharing links to their social media and related content, hinting at a broader conversation about AI's current limitations.
- GSM-Symbolic: Understanding the Limitations of Mathematical Reasoning in Large Language Modelserformance on instruction-based tasks, resulting in more accurate and contextually relevant completions. - *“This can result in improved ability to generate text that aligns closely with prompts.”* - **Risk of overfitting in narrow training**: There are potential risks of **overfitting** if the training dataset for the text completion task is not diverse or is too small, leading to high performance on familiar data but poor generalization. - A caution emphasized was that **performance might degrade** on other tasks not seen during the specific training. - **Adaptation to new formats and fields**: The training can help the model adapt to new formats or types of instructions that it was not exposed to during its initial training. - This adaptation fosters the model's ability to handle a broader range of inputs more effectively. - **Importance of diverse training datasets**: To achieve optimal outcomes when training instruct models, using a diverse and representative dataset is essential. - Monitoring for overfitting and ensuring broad performance evaluation across tasks is crucial for maintaining generalization capabilities. **Link mentioned**: : Recent advancements in Large Language Models (LLMs) have sparked interest in their formal reasoning capabilities, particularly in mathematics. The GSM8K benchmark is widely used to assess the mathemat... of overfitting in narrow training**: There are potential risks of **overfitting** if the training dataset for the text completion task is not diverse or is too small, leading to high performance on familiar data but poor generalization. - A caution emphasized was that **performance might degrade** on other tasks not seen during the specific training. - **Adaptation to new formats and fields**: The training can help the model adapt to new formats or types of instructions that it was not exposed to during its initial training. - This adaptation fosters the model's ability to handle a broader range of inputs more effectively. - **Importance of diverse training datasets**: To achieve optimal outcomes when training instruct models, using a diverse and representative dataset is essential. - Monitoring for overfitting and ensuring broad performance evaluation across tasks is crucial for maintaining generalization capabilities. **Link mentioned**:
- Tweet from Ky⨋ Gom⨋z (U/ACC) (HIRING) (@KyeGomezB)sharing` >: The Swarms framework was the first ever production-grade multi-agent orchestration framework. OpenAI has stolen our name, code, and methodology. Everything down from the syntax of the agent structure ...e config from @palebluedot_32091, who confirmed its accuracy but mentioned potential improvements. - *At glance it looks correct*, said @le_mess, suggesting enabling **sample packing** in the config. - **Importance of sample packing**: Member @le_mess highlighted the need to enable **sample packing** while running the training command `accelerate launch -m axolotl.cli.train config.yml`. - They also mentioned challenges with **multi-GPU** setups and using **LoRA** in earlier discussions. - **Fine-tuning with existing adapter**: Member @vortexwault inquired about refining a fine-tuned model using **Llama 3** as a base with an adapter for better accuracy. - @palebluedot_32091 advised merging the adapter into a new base model before training on additional data. - **Guide for merging adapters**: When @vortexwault asked for guidance on merging the adapter, @palebluedot_32091 directed them to a [GitHub guide](https://github.com/axolotl-ai-cloud/axolotl?tab=readme-ov-file#merge-lora-to-base). - He also indicated that the new training run's config should include the line `base_model: '/path/to/merged/model'`. - **Follow-up on guidance**: @vortexwault thanked @palebluedot_32091 for the guidance and mentioned he would look into the recommended resource. - "Let me check it brother thank you," said @vortexwault, indicating his intent to follow the advice.
- DisTrO and the Quest for Community-Trained AI Models | Andreessen Horowitz25574913376318)** (1 messages): >: Bowen Peng and Jeffrey Quesnelle of Nous Research discuss their mission to accelerate open source AI research, including with a new project called DisTrO. Launches for Node.js Enthusiasts**: Introducing [Swarm.js](https://github.com/youseai/openai-swarm-node), a lightweight Node.js SDK for orchestrating **multi-agent systems** using the **OpenAI API**. - This framework, inspired by the original Python version, allows developers to easily create and manage agents that perform tasks and communicate with each other. - **Key Features of Swarm.js Unveiled**: Swarm.js boasts features like **simple orchestration** of agents, custom instructions, and a lightweight structure designed for **real-world applications**. - Its ergonomic design aims to enhance usability and flexibility for developers looking to harness OpenAI's capabilities. - **Installation and Getting Started**: To install Swarm.js, simply run `npm install openai-swarm-node` and begin integrating with the OpenAI API. - The launch encourages both beginners and experts to jump in and explore the framework's potential in their projects. - **Community Contributions Are Welcome!**: The project lead invites users to contribute by checking out the [GitHub repo](https://github.com/youseai/openai-swarm-node), raising issues, and submitting pull requests. - This collaborative approach aims to enrich the SDK and foster a vibrant community around Swarm.js. **Link mentioned**:
aider (Paul Gauthier) ▷ #links (3 messages):
: aider is AI pair programming in your terminal
Nous Research AI ▷ #general (208 messages🔥🔥):
Links mentionedsily**: To get started, simply run `npm install openai-swarm-node` to install **Swarm.js**. This lightweight SDK is designed for practical, real-world use cases. - The system encourages developers of all levels to contribute to the project, promoting community involvement. - **Call for Contributions on GitHub**: The project is open for contributions, inviting users to check the [GitHub Repo](https://github.com/youseai/openai-swarm-node) to raise issues and submit pull requests. All developers, whether beginners or experts, are welcome to help enhance the framework. - Active collaboration and idea sharing are encouraged to build something outstanding together in the **open-source** community. **Link mentioned**: : - Supported languagess in an out-of-resources error with shared memory. - It includes a minimal reproduction code snippet and encourages collaboration to resolve the identified problem. **Link mentioned**: : Aider supports pretty much all popular coding languages.ludes a minimal reproduction code snippet and encourages collaboration to resolve the identified problem. **Link mentioned**:
- Tweet from Toby Pohlen (@TobyPhln) an appropriate `score_mod` function to be defined. - It's suggested that a **tradeoff** for using Flex is acceptable since it simplifies the process, despite potential reliance on CUDA with high compute capabilities. - **Logit Softcapping and Flex APIs**: It was noted that while **logit softcapping** isn't directly available with `F.sdpa`, it can be exposed in `_sdpa_or_flex_attention()` through Flex APIs, allowing for its implementation. - The discussion highlighted that shifting to **Flex** could allow quicker adaptation for upcoming attention operators in future models. - **Testing and Implementation Concerns**: Concerns were raised about the **testing** required for Gemma 2 across different model sizes, with a suggestion to crowdsource testing efforts for those with more computational resources. - Participants agreed on the importance of ensuring that the **implementation** is robust even if it means relying on a simpler fallback where needed. - **CUDA Dependency Considerations**: There was a consensus that the implementation of **Gemma 2** via Flex may be limited by requiring **CUDA** with sufficient compute capability, highlighting a gap for non-CUDA users. - Members acknowledged that this restriction would allow for more focused efforts to develop the support features needed incrementally. - **Naive Fallback vs Flex**: A member proposed that having a **`naive_fallback`** could be beneficial to support all functionalities as they emerge, despite the possible overhead in maintenance. - Concerns were shared about whether Flex's flexibility might alleviate these maintenance challenges over time.