[AINews] not much happened today
This is AI News! an MVP of a service that goes thru all AI discords/Twitters/reddits and summarizes what people are talking about, so that you can keep up without the fatigue. Signing up here opts you in to the real thing when we launch it 🔜
a quiet weekend is all you need.
AI News for 10/24/2024-10/25/2024. We checked 7 subreddits, 433 Twitters and 32 Discords (232 channels, and 3136 messages) for you. Estimated reading time saved (at 200wpm): 319 minutes. You can now tag @smol_ai for AINews discussions!
- Liquid AI held a launch event (our coverage here)
- Anthropic shared some social bias study followups on "Golden Gate Claude" feature steering
- Cohere followed up Aya Expanse with multimodal Embed 3 embeddings models.
- there was some fake news on GPT5/Orion.
Happy weekend.
The Table of Contents and Channel Summaries have been moved to the web version of this email: !
AI Twitter Recap
all recaps done by Claude 3.5 Sonnet, best of 4 runs.
AI Models and Research
- Meta FAIR’s Open Materials 2024: @AIatMeta announced the release of Open Materials 2024, featuring new models and datasets for inorganic materials discovery, utilizing the EquiformerV2 architecture and supporting extensive data on structural and compositional diversity.
- Anthropic AI’s Feature Steering: @AnthropicAI shared their research on feature steering, demonstrating how adjusting model features can influence social bias scores across nine dimensions, while identifying a "steering sweet spot" that balances effectiveness and capability retention.
- NVIDIA’s Llama-3.1-Nemotron-70B: @lmarena_ai revealed that Llama-3.1-Nemotron-70B now ranks #9 and #26 on the Arena leaderboard with Style Control, showcasing its competitiveness in human preference tasks.
- Perplexity’s Model Enhancements: @AravSrinivas highlighted Perplexity’s growth to over 100M weekly queries and introduced new features like Finance and Reasoning Mode, enhancing its capabilities and user engagement.
AI Tools and Infrastructure
- LangChain’s Application Integration: @hwchase17 emphasized the integration of LangChain into real applications, supporting features like Interactive Frame Interpolation to enhance deployment scenarios.
- Kestra’s Event-Driven Workflows: @svpino discussed adopting Kestra for scalable, event-driven workflow orchestration, highlighting its open-source nature, YAML-based workflows, and ability to handle millions of executions.
- OpenFLUX Optimization: @ostrisai explored training a guidance LoRA with OpenFLUX to double inference speed by eliminating CFG, showcasing practical optimizations for AI models.
AI Safety and Ethics
- Trust in Humans vs. AIs: @RichardMCNgo illustrated how trust dynamics differ between humans and AIs, emphasizing the importance of human oversight in AI-driven research to ensure reliability and prevent misuse.
- Economic and Intellectual Impact of AI: @ajeya_cotra and @tamaybes discussed the profound economic transformations driven by AI automation, predicting significant growth rates and highlighting the critical role of human intelligence in verifying AI-generated findings.
- White House AI National Security Memo: @DanHendrycks shared insights from the White House’s AI National Security strategy, focusing on mitigating AI risks in offensive cyber operations and biological threats, underscoring the importance of national security measures in AI deployment.
AI Applications and Use Cases
- LlamaIndex’s Knowledge-Backed Agents: @jerryjliu0 presented how LlamaIndex workflows enhance AI agent applications by incorporating event-driven architecture and robust state management for improved performance and reliability.
- Perplexity’s Financial Search API: @virattt introduced a new Financial Search API that enables searching across 20,000 tickers with over 100 filters, streamlining financial data processing and analysis for users.
- AI Agents in Sales Automation: @llama_index showcased a case study on deploying LlamaIndex for NVIDIA’s internal AI assistant for sales, detailing its use of multi-agent systems, parallel retrieval, and real-time inference to enhance sales automation and efficiency.
AI Community and Events
- AI Agents Masterclass: @jerryjliu0 conducted an AI Agents Masterclass with @arizeai, covering the fundamentals of building knowledge-backed agents using LlamaIndex workflows, with a focus on event-driven architecture and state management.
- Podcasts and Workshops: @swyx and @maximelabonne promoted upcoming podcasts and workshops focused on AI development, community engagement, and collaborative learning, fostering a vibrant AI community.
- Meta FAIR’s Open Materials Workshop: @maximelabonne organized a workshop on Meta’s Open Materials, inviting AI researchers and enthusiasts to collaborate on inorganic materials discovery using open-source models and datasets.
Memes/Humor
- AI Takeover Jokes: @RichardMCNgo humorously likened AI submissions to those written by Einstein, suggesting a humorous scenario where AI could potentially collude to take over the world.
- Funny AI Predictions: @francoisfleuret made light-hearted remarks about AI task arithmetic and layer complexity, blending technical insights with humor.
- AI-Generated Music: @suno_ai_ shared an AI-generated song, humorously transforming a tweet into a bat gothclub music theme, showcasing creative and entertaining uses of AI in content generation.
- Humorous AI Comparisons: @teortaxesTex joked about AI’s attempt to write a paper on intelligence and order, highlighting the amusing limitations of AI-generated content.
AI Reddit Recap
/r/LocalLlama Recap
Theme 1. Meta's Quantized Llama Models: Pushing On-Device AI Forward
- Introducing quantized Llama models with increased speed and a reduced memory footprint (Score: 75, Comments: 3): Meta released quantized versions of their Llama 2 models, offering 2-3x faster inference and a 40-60% reduction in memory usage. The new models, available in 4-bit and 8-bit quantization, maintain performance comparable to their full-precision counterparts across various benchmarks, including MMLU, HellaSwag, and TruthfulQA. These quantized models aim to improve accessibility and efficiency for developers working with large language models on resource-constrained devices.
- Zuck on Threads: Releasing quantized versions of our Llama 1B and 3B on device models. Reduced model size, better memory efficiency and 3x faster for easier app development. đź’Ş (Score: 404, Comments: 103): Meta has released quantized versions of their Llama 1B and 3B on-device models, as announced by Mark Zuckerberg on Threads. These new versions offer reduced model size, improved memory efficiency, and are 3x faster than their predecessors, aimed at facilitating easier app development for developers.
- Quantization-Aware Training (QAT) with LoRA adaptors was used for the new models, involving multiple training steps to achieve high-quality post-quantization results. This process is difficult for the open-source community to replicate due to dataset quality and format uncertainties.
- The quantization scheme includes 4-bit groupwise quantization for linear layers in transformer blocks, 8-bit per-channel quantization for classification and embedding layers, and uses PyTorch's ExecuTorch framework with Arm CPU backend in mind.
- Users discussed the importance of official model sources for businesses, with some expressing challenges in using models like Qwen 2.5 due to their Chinese origin, particularly in defense contracting contexts.
- Meta released quantized Llama models (Score: 184, Comments: 25): Meta released quantized Llama models using Quantization-Aware Training, LoRA, and SpinQuant techniques, marking their first release of such versions. These models demonstrate impressive performance despite significant size reductions, making them suitable for widespread deployment due to their small size and fast speed; they can be accessed and used via executorch on GitHub.
- QLoRA variants show impressive results, with users discussing similarities to the QLoRA method from Tim Dettmers' paper. Questions arose about the use of QLoRA in popular quantization methods and its dependence on compute power.
- Most post-training quantization methods (e.g., Q5_0 gguf) don't include a LoRA component. Meta's approach using original datasets and early-stage training leads to higher accuracy than typical open-source PTQ models.
- Users inquired about converting the models to GGUF format for use in LM Studio, with discussions noting these smaller models are more suited for devices like phones rather than Macs. Interest was also expressed in potential 128k context length models for applications like Skyrim role-playing.
Theme 2. Cerebras Inference Achieves 2,100 Tokens/s on Llama 3.1-70B
- Cerebras Inference now 3x faster: Llama3.1-70B breaks 2,100 tokens/s (Score: 214, Comments: 81): Cerebras Inference has achieved a 3x performance boost, now running Llama 3.1-70B at 2,100 tokens per second. This performance is 16x faster than the fastest GPU solution and 8x faster than GPUs running Llama3.1-3B, a model 23x smaller, with the improvement comparable to a new GPU generation upgrade. Companies like Tavus and GSK are using Cerebras Inference for video generation and drug discovery, with a chat demo and API available at inference.cerebras.ai.
- The Cerebras CS-2 hardware is a 15U machine with a 23kW power draw, costing around $1-3 million. It features 40GB of on-chip SRAM and uses entire pizza-sized wafers from TSMC instead of cut chips. A server teardown video showcases its unique architecture.
- Users report impressive performance on the Cerebras chat demo, particularly for translation tasks. The demo runs Llama 3.1 70B & 8B models, with some users finding it superior to OpenAI's offerings. However, concerns were raised about API usage limits and first token latency.
- Discussions touched on potential applications, including scaled thinking for o1-like models, inference-time compute scaling, and better samplers. Some users questioned the comparison metrics, suggesting a need for standardized measurements like "watts per million tokens" for fair hardware comparisons.
Theme 3. New Open Source LLMs Push Boundaries of Context Length and Capabilities
- INTELLECT-1: groundbreaking democratized 10-billion-parameter AI language model launched by Prime Intellect AI this month (Score: 170, Comments: 37): Prime Intellect AI has released INTELLECT-1, a 10-billion-parameter AI language model, marking a significant advancement in democratized AI technology. The model, launched this month, aims to provide accessible and powerful language processing capabilities to a wider range of users and developers, potentially reshaping the landscape of AI applications and research.
- CohereForAI/aya-expanse-32b · Hugging Face (Context length: 128K) (Score: 145, Comments: 57): CohereForAI has released Aya Expanse 32B, a large language model with a 128K token context length, available on Hugging Face. This model represents a significant advancement in context handling capacity, potentially enabling more comprehensive and contextually aware language processing for various applications.
- Users expressed skepticism about the model's performance, with many calling for comparisons to Qwen 2.5. Some noted that US and European companies seem to be ignoring Qwen's achievements, despite its better license and output in certain use cases.
- There was discussion about a potential config mistake in the model, as the
max_position_embeddingsvalue (8192) doesn't match the stated 128K token context length. This issue was similar to a previous release from CohereForAI, as discussed in a Hugging Face thread. - The 8B version of the model was tested and found to be highly aligned and moralizing, refusing seemingly mundane requests. Users noted that the model's primary purpose is for translation tasks, not general use, and a q8 gguf version was made available on Hugging Face.
Theme 4. Improving LLM Integration for Developers and Mobile Users
- VSCode + Cline + VLLM + Qwen2.5 = Fast (Score: 99, Comments: 29): The post describes an integration of VSCode, Cline, VLLM, and Qwen2.5 for rapid coding assistance. This setup allows for fast local AI-powered code completion and generation, leveraging the speed of VLLM and the capabilities of the Qwen2.5 model within the VSCode environment.
- ChatterUI v0.8.0 released - Now with external model loading! (Score: 35, Comments: 13): ChatterUI v0.8.0, an Android UI for LLMs, has been released with significant updates, including external model loading capability. The app now separates Remote and Local modes, with Local Mode allowing users to customize and use on-device models, while Remote Mode enables connection to various supported APIs. Key improvements include a new model list inspired by Pocket Pal, displaying metadata extracted from GGUF files, and a Model Settings Page with CPU settings and local-specific app options.
Theme 5. Advancements in LLM Benchmarking and Evaluation Tools
- Benchmark GGUF models with ONE line of code (Score: 45, Comments: 20): The post introduces an open-source tool for benchmarking GGUF models with a single line of code, addressing challenges in evaluating quantized models locally. The tool supports multiprocessing, 8 evaluation tasks, and is claimed to be the fastest benchmark for GGUF models, with an example showing a Llama3.2-1B-Instruct Q4_K_M model evaluation taking 80 minutes on a 4090 GPU using 4 workers for the "ifeval" dataset.
- Users expressed interest in testing custom models without uploading them, particularly for comparing static vs imatrix quantizations. The tool's flexibility for evaluating various model types was highlighted.
- A question was raised about the possibility of measuring power consumption and efficiency for specific models on devices like the MacBook Pro M1, indicating interest in performance metrics beyond speed.
- Enthusiasm was shown for testing the benchmark tool on different hardware, including AMD Ryzen GPUs, suggesting a desire for broader compatibility and performance comparisons across various GPU architectures.
- Power scaling tests with 4X RTX 3090's using MLC LLM and Mistral Large Instruct 2407 q4f16_1. Tested 150 - 350 watts. (Score: 44, Comments: 23): Power scaling tests were conducted using 4 RTX 3090 GPUs with MLC LLM and Mistral Large Instruct 2407 q4f16_1, exploring a power range of 150 to 350 watts. The experiments aimed to evaluate the performance and efficiency of these high-end GPUs in running large language models at various power levels.
- SuperChewbacca used the prompt "Write exactly 100 digits of pi" for testing, running MLC LLM in chat mode with tensor parallel shards=4. They appreciated MLC LLM's speed and consistent 100% GPU utilization.
- Users expressed interest in comparing MLC LLM's performance to vLLM for Mistral-large, particularly regarding tensor parallelism efficiency. The original poster agreed to conduct a comparable quantization test in vLLM.
- Requests were made to include Ollama and vLLM in future benchmarks for a comprehensive tok/s comparison across all three solutions on the 4x3090 setup.
Other AI Subreddit Recap
r/machinelearning, r/openai, r/stablediffusion, r/ArtificialInteligence, /r/LLMDevs, /r/Singularity
AI Model Releases and Capabilities
- Mochi 1 video generation model: A new open-source AI model called Mochi 1 demonstrates impressive video generation capabilities. It can run on a single 24GB VRAM GPU card with some optimizations. The model can generate 15 second videos at 24 fps in fp8 precision or 2.5 second videos in bf16 precision. A detailed guide was shared on how to set it up and run it locally.
- Anthropic's Claude 3.5 models: Anthropic released new Claude 3.5 models with a "computer use" capability, allowing the AI to directly interact with computer interfaces. This is seen as a major step towards AI agents that can automate computer-based tasks. The release sparked discussion about its potential impact on knowledge work and automation.
- OpenAI's next model: There were conflicting reports about OpenAI's plans to release a new AI model codenamed "Orion" by December. While some sources reported this, OpenAI CEO Sam Altman dismissed it as "fake news". The conflicting information led to much speculation in the AI community.
AI Research and Techniques
- Google DeepMind's multimodal learning: A new paper from Google DeepMind demonstrates how data curation via joint example selection can accelerate multimodal learning.
- Microsoft's MInference: Microsoft introduced MInference, a technique that enables inference of up to millions of tokens for long-context tasks while maintaining accuracy and dramatically speeding up supported models.
- Scaling synthetic data creation: A paper on scaling synthetic data creation leverages diverse perspectives within large language models to generate data from 1 billion personas curated from web data.
AI Model Improvements
- Salesforce's xLAM-1b model: Salesforce released xLAM-1b, a 1 billion parameter model that achieves 70% accuracy in function calling, surpassing GPT 3.5 despite its relatively small size.
- Phi-3 Mini update: Rubra AI released an updated Phi-3 Mini model with function calling capabilities, competitive with Mistral-7b v3 and outperforming the base Phi-3 Mini.
AI Ethics and Societal Impact
- The rapid advancement of AI capabilities, particularly in automating computer-based tasks, sparked discussions about potential job displacement and the need for solutions like Universal Basic Income (UBI).
- There were debates about the concentration of AI power in the hands of a few companies, with some criticizing OpenAI's apparent shift away from its initial open-source charter.
Hardware and Infrastructure
- TSMC's Arizona chip production yields reportedly surpassed those in Taiwan, seen as a win for US semiconductor manufacturing efforts.
AI Discord Recap
A summary of Summaries of Summaries by O1-preview
Theme 1. AI Models and Hardware Break New Ground
- Cerebras Chip Leaves GPUs in the Dust: Cerebras unveils a chip delivering 3x faster inference, achieving over 2,100 tokens/s with Llama3.1-70B, outpacing the fastest GPUs by 16x. This leap positions Cerebras as a heavyweight in AI processing speed.
- Intel Arc A750 Surprises Everyone: Upgrading to the Intel Arc A750, users found impressive performance in LM Studio, surpassing previous setups like the 6750xt. This highlights the Arc's potential in machine learning tasks.
- Meta Releases Lightning-Fast Quantized Llama Models: Meta drops quantized versions of Llama 3.2 1B & 3B, boosting inference speed by up to 4x. Aimed at on-device deployments, these models balance speed and performance.
Theme 2. Ethical Challenges and Privacy Concerns in AI
- Claude 3.5 Becomes Big Brother: The new Claude 3.5 Sonnet can monitor screens and control devices, raising serious privacy red flags. Users debate the ethics of AI with such intrusive capabilities.
- Deepfake Tech Gets Too Real for Comfort: On Notebook LM, discussions heat up over deepfake technology's ethical implications, especially concerning consent and dehumanization. Members question whether AI-generated avatars can ever be ethical.
- AI Censorship Sparks Outrage: Users on OpenRouter worry about potential censorship of models like hermes-3-llama-3.1-405b, fearing restrictions on content. The community debates where to draw the line on acceptable AI content moderation.
Theme 3. User Experiences with AI Tools and Platforms
- LM Studio Users Demand Plugins NOW!: A chorus of users calls for LM Studio to support user-created plugins, seeking enhanced functionality without added complexity. Better integration with existing tools and APIs is a hot topic.
- Aider Gets an Upgrade, Users Cheer: The release of Aider v0.60.1 brings support for Claude 3 models, file sorting, and a fancy new input flag. Users appreciate the updates, noting improvements in cost savings through prompt caching.
- Perplexity Pro Divides the Crowd: The launch of Perplexity Pro spurs debate over its value against competitors like Claude and GPT. Users question performance versus price, seeking advice on optimizing their subscriptions.
Theme 4. AI-Assisted Creativity Takes Center Stage
- AI Podcasting Gets Personal and Weird: On Notebook LM, users find that assigning names and roles to AI voices enhances coherence in generated podcasts. However, limitations in voice roles spark creative challenges.
- AI Quiz Game Fumbles the Score: Attempts to create an AI-powered quiz game see initial success but stumble when the AI can't tally scores. The AI's notorious math struggles become a playful topic among users.
- Writers Level Up with AI Co-Authors: Authors use AI to flesh out characters and scenes, finding 'table reads' with AI deepen narratives. This method uncovers new backstories and motivations, boosting creative writing.
Theme 5. Fine-Tuning AI: Challenges and Best Practices
- Bad Data In, Bad AI Out: Dataset Quality Matters: Unsloth AI users emphasize that fine-tuning success hinges on high-quality, balanced datasets. Unbalanced data leads to poor performance, underscoring the need for proper preparation.
- Fine-Tuning Llama 3.2 Sparks Debate: On Eleuther, members discuss the best approaches to fine-tune Llama 3.2 for text classification. Suggestions include using simple classifiers and embedding models, with caution about dataset quality.
- Quantization Techniques Raise Eyebrows: In Nous Research AI, the community examines Meta's new quantized models, weighing the benefits and complexities of applying quantization-aware training. Potential performance trade-offs spark lively debates.
PART 1: High level Discord summaries
HuggingFace Discord
-
H200 Servers Crush AI Model Performance: A discussion centered on using H200 servers to run large models revealed that one user’s production server was processing 405B models at 90 teraflops.
- Concerns arose regarding the cost-effectiveness and necessity of such robust infrastructure for typical AI applications.
- Transformers Basics and Reddit Generation: A member shared their progress in learning transformers, leveraging Andrej's video to achieve results with a 10M parameter model, generating 10k tokens from Reddit data.
- This milestone sparked a discussion on further optimizations and community feedback for their DeepLLMs repository.
- Automated Penetration Testing Benchmark Introduced: The paper “Towards Automated Penetration Testing: Introducing LLM Benchmark, Analysis, and Improvements” highlights a benchmark geared towards using LLMs for penetration testing, evaluating GPT-4o and Llama 3.1.
- With cyber threats costing $6 trillion, the discussion emphasized the necessity for ethical hacking and benchmarks for effective vulnerability identification.
- Streamlit Calculator Project Unveiled: A member replicated a Calculator project using Streamlit, inviting feedback on their implementation.
- The excitement around this project complemented discussions on utilizing HuggingFace tools for protein analysis in genomics.
- Contributions to Hugging Face Diffusers Explored: Interest in contributing to Hugging Face Diffusers led to recommendations for reading the contributing readme and identifying good first issues.
- As discussions unfolded, queries regarding the impact of adding noise to tensors without retraining arose, highlighting community engagement in technical challenges.
Unsloth AI (Daniel Han) Discord
-
Unsloth AI progresses with model support: Unsloth currently lacks vision model support like Llama 3.2, but the team is developing capabilities to include them in the future.
- Users are urged to focus on text-based LLMs while the integration of vision models is in the works.
- Finetuning model for subtitles meets challenges: A user reports difficulties finetuning a model to correct VTT subtitles, with issues stemming from timestamp alterations during training.
- Experts recommend removing timestamps from training datasets to avoid overfitting and enhance text correction capabilities.
- Quality of datasets is paramount for finetuning: The success of LLM finetuning hinges on the quality and balance of the training dataset, with unbalanced data leading to subpar performance.
- Participants emphasized the importance of proper dataset preparation prior to training.
- Data centers booming with a 180% increase: Discussion surfaced around a staggering 180% increase in data center construction in 2024, potentially marking a significant trend in the sector.
- Some members expressed skepticism, suggesting it may simply indicate wasted investments rather than a sustainable growth trajectory.
- Nvidia's stronghold in the AI domain: Debate on Nvidia's market share reflects on its historical reliance on gaming, transitioning now to focus on AI accelerators.
- One member asserted that enterprises would still prefer Nvidia over AMD, even if the latter’s offerings were free, highlighting brand loyalty.
Latent Space Discord
-
E2B Desktop Sandbox Launches: The E2B Desktop Sandbox is now in beta, creating isolated environments tailored for LLM applications, featuring full filesystem support and robust customizability.
- User feedback is encouraged to refine the platform and optimize its utility in cloud environments.
- Claude 3.5 Pushes Privacy Boundaries: The new Claude 3.5 Sonnet can now monitor screens and control devices, offering capabilities like file searching and web automation that raise significant privacy concerns.
- This advancement highlights a substantial leap in AI interaction complexity, provoking discussions about ethical usage.
- Cerebras Chip Sets New Inference Records: A new chip from Cerebras demonstrates 3x faster inference performance with Llama3.1-70B, achieving over 2,100 tokens/s, outpacing the fastest GPUs by 16x.
- This breakthrough positions Cerebras as a significant player in the AI processing landscape, setting a high benchmark for competitors.
- OpenAI's Orion Speculation Creates Buzz: OpenAI hints at launching a model called Orion by December, sparking debates amid accusations of misinformation regarding its development timeline.
- CEO Sam Altman’s remarks on forthcoming technologies are stirring speculation and confusion about the actual release schedule.
- Cohere's Embed 3 Enhances Multimodal Capabilities: Cohere introduced its Embed 3 model, which allows enterprises to conduct searches across both text and image datasets, vastly improving AI functionality.
- This development aims to facilitate real-time data processing across diverse document types, fostering greater efficiency.
Notebook LM Discord Discord
-
Podcast Customization Enhances Coherence: Users have discovered that specific prompts, such as assigning names and roles, enable cohesion in AI-generated podcasts, keeping host introductions consistent throughout episodes.
- Roles limitation became evident as the male voice typically plays the host while the female voice acts as an expert, complicating flexibility in casting.
- Deepfake Technology Sparks Ethical Concerns: Discussion on deepfake technology raised ethical issues surrounding consent, emphasizing its crucial role in public understanding to avoid potential misuse.
- Members worried about dehumanization in AI, questioning if avatars could be ethically produced while suggesting that responsibility falls on content creators.
- AI Quiz Game Developments Underway: Users trialed a quiz game format using AI for dynamic question exchanges, spotting initial success but trailing off in accurately counting scores.
- The discrepancies in tallying responses highlighted persistent challenges, notably AI's shortcomings in mathematical accuracy.
- Character Development with AI Assistance: Utilizing AI, members are examining screenplay drafts for character gaps and development, leading to improved storylines through 'table reads'.
- This method has produced deeper narrative insights and potential backstory ideas via more engaging interactions with the AI.
- AI Performance Limitations Expose Weaknesses: Participants acknowledged the AI's tendency to hallucinate, particularly in counting and factual delivery, greatly impacting overall accuracy.
- Discussions included leveraging additional tools like Python to overcome these computational shortcomings in AI.
LM Studio Discord
-
Users Crave LM Studio Plugin Features: There’s growing interest in the potential for user-created plugins in LM Studio, which could enhance functionality without adding complexity.
- Better integration with existing tools and open API endpoints could significantly improve user experience.
- Mamba-Codestral Model Fails to Load: A user reported issues loading the Mamba-Codestral model, hinting at GPU errors and driver conflicts as the primary culprits.
- Suggested fixes included cleaning shader caches and modifying GPU offload percentages to address VRAM limitations.
- Performance of Large Language Models Reviewed: Users discussed experiences with large LLMs, noting that bigger sizes can enhance context length but escalate hardware demands.
- Performance optimization remains a concern as larger models can slow down response times due to resource strain.
- Intel Arc A750 Surprises with Speed: After upgrading to the Intel Arc A750, a user found impressive performance in LM Studio, outpacing their previous 6750xt setup.
- This sparked conversations about the capabilities of modern GPUs, especially in machine learning contexts.
- Gemma 2 Tokens Rates and Concerns: The Gemma 2 2B model reached 25 tokens/s, while Gemma 2 9B lagged at approximately 6 tokens/s, raising flags about output errors.
- These token speeds highlight issues that may impede model usability, necessitating further investigation.
aider (Paul Gauthier) Discord
-
DeepSeek delivers quick performance: While using DeepSeek for the editor-model, users noted no significant slowdown during processing, prompting excitement about the tool's efficiency.
- This positive feedback indicates potential in adopting DeepSeek for smoother coding experiences.
- What's new in Aider v0.60.1: Upcoming Aider v0.60.1 includes support for Claude 3 models, file sorting, and a new --fancy-input flag to enhance command handling.
- Speculations arose regarding the delay in installation, hinting at local issues that some users encountered.
- Prompt caching saves costs: Users explored prompt caching options in Aider, finding it beneficial for enhancing performance and reducing costs, especially with the Sonnet model.
- Enabling caching reportedly minimizes expenses tied to local coding tasks, making it a preferred tactic.
- PearAI integrates Aider: Discussion emerged around PearAI using Aider for coding features, leading to questions about permissions and the nature of the integration.
- Concerns surfaced regarding possible rebranding or alteration of Aider’s capabilities within PearAI, noted in the PearAI Creator article.
- Concerns over Claude 1022 behavior: Users reported unpredictable outputs from Claude 1022, often citing 'hyperactive' behavior when working with tools like Cursor.
- The inconsistency in output led to discussions on needing refined user prompts to maintain control during interactions.
Nous Research AI Discord
-
Nous Research Secures Revenue Sharing: Nous Research partners with Hyperbolic to share revenue from their Hermes 3 model, fostering a collaborative funding approach.
- Members discussed the partnership as a mutually beneficial arrangement, clarifying it’s not a case of 'selling out'.
- Diminishing AI Hype Cycle: Members noted a decrease in AI hype compared to earlier in the year, potentially overshadowed by events like the upcoming US elections.
- The conversation speculated that the community might be in a phase of 'inflated expectations' rather than true engagement.
- Benchmarking Model Performance: A lively debate occurred over the Llama 4 model's performance compared to Claude, with skepticism about current benchmark methods.
- One member pointed out Llama 4's performance at 120+ tps, questioning the validity of the comparisons.
- Exploring Quantization Techniques: Members analyzed the introduction of quantized models by Meta, debating their feasibility and potential benefits for model training.
- Concerns were raised regarding the computational complexity associated with applying quantization-aware training.
- Softmax Function Under Investigation: A paper from Google DeepMind reveals that the softmax function struggles with sharpness as inputs grow, leading to dispersed attention coefficients.
- Experiments indicate that while models excel with familiar tasks, their focus weakens in larger, out-of-distribution cases.
Eleuther Discord
-
NEO Tests Show Improvement: Local testing of the NEO model reveals it is becoming faster and smarter with repeated interactions, sparking interest in the training pile.
- Commenters noted the engaging nature of interactions in these tests.
- Munkres Recommended for Topology: In a request for good topology books, members quickly recommended Munkres as a reputable source for study.
- This book has gained a strong reputation among topology students.
- Fine-Tuning Llama 3.2 Model: A member sought guidance on fine-tuning the Llama 3.2 model to categorize text into 20 categories, specifically about the usage of DPO.
- Suggestions involved employing simple classifiers, although members raised concerns about potential performance issues with the dataset.
- Classifier-Free Guidance Doubts: Skepticism arose about the effectiveness of Classifier-Free Guidance (CFG), pointing out issues with its dependence on timestep and guidance scales.
- The conversation included a potential simplified approach to generating outputs directly from textual input.
- Challenges with Image Captioning Datasets: Concerns were raised regarding the poor quality of captions in datasets, with claims that re-captioning won't resolve accuracy and relevance issues.
- The challenge of generating high-quality captions at scale was debated, underscoring the limitations of existing solutions.
OpenAI Discord
-
Opus 3.5 Release Faces Timeline Uncertainty: Speculation arises about whether Opus 3.5 from Anthropic will launch this year, with some believing it may be delayed until 2025.
- It's suggested that they might be jumping straight to a newer version instead.
- AGI vs ANI Debate Heats Up: Members engage in a spirited discussion debating Artificial Narrow Intelligence (ANI) versus Artificial General Intelligence (AGI), assessing the definitions and applicability of these terms to current AI models.
- Some propose using the term Emerging AGI to describe potential pathways toward developing general intelligence.
- Future AI Training Methods Speculated: Discussion centers around the resources needed for models operating at the scale of millions of H100s, raising concerns about production issues with next-gen GPUs.
- Achieving this scaling may still depend heavily on existing hardware with some estimating significant requirements ahead.
- OpenAI's Data Center Ambitions Spark Debate: A recent report outlines OpenAI's plans to build extensive 5GW data centers for training advanced AI models, launching conversations about feasibility and scale.
- Skeptics worry about the ecological impact and practicality of such extensive compute goals.
- Co-Pilot Icon MIA After Update: A user experiences the Co-Pilot icon vanishing from their Windows system following an update, prompting inquiries into the cause and possible fixes.
- Responses range from confusion to jokes, revealing a shared user experience across the community.
OpenRouter (Alex Atallah) Discord
-
Cerebras API Access Sparks Interest: Users shared their experiences with Cerebras API, noting varied timelines for access ranging from over a month ago to acquiring keys without formal acceptance.
- Discussions highlighted the balance between chip costs and the expected performance from the API.
- Censoring Speculation on Hermes-3: Concerns were raised regarding potential censorship of hermes-3-llama-3.1-405b, reflecting community worries about model content restrictions.
- This points to a larger conversation about the thresholds for acceptable content in AI models.
- Exploring Prompt Caching Benefits: The availability of prompt caching for Sonnet models on OpenRouter was discussed, with users emphasizing its ability to optimize API usage.
- However, some encountered implementation issues, particularly when interfacing with external applications like SillyTavern.
- Token Limits Frustrate Users: Frustrations emerged over API token limits, after one user received a max tokens limit error despite having $16 in credits, leading to discussions about creating new API keys.
- The community consensus leaned towards checking account credit status as part of troubleshooting.
- Performance Concerns with OpenRouter: Users reported encountering slowdowns and error 520, raising alarms about system reliability and performance issues.
- The discussion pointed out that hardware supply challenges are impacting the performance of advanced models.
Stability.ai (Stable Diffusion) Discord
-
Flux Faces Comic Creation Challenges: Members discussed using FLUX for comic generation, highlighting the need for specific character model fine-tuning to enhance consistency and prompt fidelity.
- It's difficult to achieve the desired level of detail with standard models, necessitating further training for specific character consistency.
- Mochi Outperforms in Video Generation: Users pitted Mochi 1 against CogVideoX for local video creation, concluding that while Mochi is superior, it has slower processing times.
- Users recommended CogVideoX for its feature set despite being less effective in certain scenarios than Mochi.
- Skepticism Around Stable Diffusion 3.5: There were questions about the abilities of Stable Diffusion 3.5 to generate targeted prompts like 'A woman lying on top of a pool of marshmallows'.
- One user noted that images created with this prompt had surfaced in another channel for community feedback.
- Artwork Creation for House Music: A member looked for tips on designing cover artwork for a house track on SoundCloud, sharing specific expectations for the artwork's style.
- Disappointment with initial results surfaced, indicating the learning curve in mastering AI-driven art generation.
- LoRA Training Relies on Good Datasets: A discussion ensued about the importance of quality datasets for LoRA model training, ensuring reliable outputs.
- Participants suggested that tutorials on dataset preparation could greatly improve user proficiency before model fine-tuning.
Perplexity AI Discord
-
Perplexity Pro Sparks User Debate: Users shared diverse experiences with Perplexity Pro, debating its value against competitors like Claude and GPT. They sought effective setups and resources to optimize their subscriptions.
- Concerns about performance versus value emerged, prompting further exploration of best use cases.
- Gemini 2.0 Release is on the Horizon: The launch of Gemini 2.0 is expected soon as Google and OpenAI race to unveil next-gen models, amidst questions on the expected performance gains. December is shaping up to be significant in AI developments.
- Participants noted the swift progress in AI capabilities but pointed out that improvements remain fragmented across different platforms.
- Inquiries on Perplexity App Features: Curiosity peaked around the Perplexity app's reasoning capabilities and its requirement for iOS speech recognition. Discussions emphasized the significance of managing instruction settings to minimize AI hallucinations.
- Users expressed concerns about ensuring reliable outputs from the app for more critical workflows.
- Legal Sector Leverages AI: Frustrations were vocalized regarding AI's role in legal research, highlighting struggles to produce reliable outputs despite meticulous prompt instructions. The need for dependable information sourcing was stressed in the discussions.
- Users exchanged techniques to refine prompts aiming to optimize AI performance in legal scenarios.
- Bitcoin Creator Mystery Unraveled: A shocking revelation has emerged about the identity of Bitcoin's creator, igniting discussions in crypto communities. The findings can be viewed in this YouTube video.
- This breakthrough could reshape conversations about Bitcoin's origins in blockchain discourse.
GPU MODE Discord
-
Exploring AI in Veterinary Medicine: A member inquired about promising applications of AI in veterinary medicine, sparking interest in innovative uses.
- This led to an open forum discussion without specific references, highlighting the untapped potential in the field.
- Triton Optimizations Show Performance Challenges: Encapsulating kernels in
custom_opresulted in a performance drop from 23 tokens/sec to 16 tokens/sec, raising concerns about the wrapping mechanism.
- Members are questioning the overhead impacts of this approach on Triton and considering further optimizations.
- Llama 3.2 Models are Now Open Source: Meta has released Llama 3.2 1B and 3B models, targeting on-device deployments and improving performance via quantization techniques.
- Developers aim to optimize memory while ensuring the models retain their effectiveness in low-resource scenarios.
- Training Enhancements for NanoGPT: Discussion highlighted that NanoGPT can gain speed via optimized Triton operations, especially if only using eager PyTorch.
- The community emphasizes the importance of incorporating torch.compile to enhance performance during model training.
- Discord Cluster Manager Development Begins: Documentation for the Discord Cluster Manager has been shared, outlining project functionality and future development needs.
- Active development is planned to commence on November 3, aiming for completion by November 10, inviting contributions from the community.
Modular (Mojo 🔥) Discord
-
General Questions Direction Clarified: Members are reminded to direct questions about the organization to the correct channel here for structured support.
- This restructuring aims to streamline inquiries, ensuring members can find answers effectively.
- Kitty Ket's LED Matrix Breakthrough: Kitty Ket reported advancements in the LED matrix project, achieving cutting-edge performance with 3D vectors and data manipulation functions.
- Processing times are targeted below 10 ms, showcasing promising results despite the absence of communication with the LED matrix.
- Integrating PostgreSQL with Mojo: A member raised a question about integrating libpq.so for PostgreSQL into Mojo, specifically regarding
ffi.external_callfor custom libraries.
- Darkmatter shed light on the translation of
char*in C, which generally converts toInt8for x86_64 andUInt8for ARM, indicating a need for clarity in integration. - New Bug Report on Mojo’s Memory Management: A recent bug report highlights that Mojo prematurely frees memory while references remain active.
- Users are unable to retain the address of a List due to immediate freeing, presenting ongoing challenges in memory management.
- Serialized Model Ingestion Use Cases Explored: Members discussed potential use cases for ingesting serialized models via the Graph API, seeking community insights.
- The engagement aims to align model ingestion development with real-world user needs and application scenarios.
tinygrad (George Hotz) Discord
-
Deterministic GPU Kernels for Metal: A member inquired about creating deterministic GPU kernels targeting Metal to achieve consistent outputs across GPUs like M2 and M3. Another highlighted that success could warrant forking tinygrad.
- This effort aligns with the broader aim of improving the consistency and reliability of GPU computations.
- Floating Point Arithmetic Consistency Challenges: Concerns emerged regarding floating-point arithmetic inconsistencies in MLX, prompting discussions about tinygrad's capacity for determinism. Users debated the implications of these inconsistencies on model reliability.
- The non-associative nature of floating-point arithmetic may present challenges in achieving consistent outputs across various environments.
- Tinygrad's Metal Configurations Revealed: Tinygrad disables Metal's fast math mode by default to mitigate discrepancies in floating point operations, driving discussions about its implications on performance. The transition to the mathMode option suggests potential pathways for improving determinism.
- Members acknowledged the importance of understanding these configurations when working on GPU-oriented projects.
- Beam Search in Kernel Space Impresses: Users expressed enthusiasm about the beam search in kernel space, noting impressive speed, albeit not matching flash attention yet. This highlights tinygrad's continued optimization capabilities.
- The discourse emphasized the effectiveness of kernel-level optimizations in accelerating search algorithms.
- Handling Environment Variables in Notebooks: A user faced challenges with setting environment variables in a notebook for the Fashion MNIST dataset, leading to confusion about necessary configurations. George Hotz clarified the proper usage of os.environ.
- This clarification helped streamline workflows, emphasizing the importance of correct environment handling in notetaking frameworks.
LlamaIndex Discord
-
Build Knowledge-Backed Agents with LlamaIndex: In an AI Agents Masterclass, the founder detailed creating a knowledge-backed agent using LlamaIndex workflows, emphasizing the LLM router and other essential tools, view the session here.
- The session compared event-based and graph-based architectures, with a consensus favoring LLM routers for their superior performance.
- NVIDIA's Internal AI Assistant Deployment: NVIDIA announced its internal AI assistant utilizing Llama 3.1 405b for simple queries and the 70b model for document searches, detailed here.
- The assistant integrates multiple information sources, including internal documents and the NVIDIA site, streamlining access to critical data.
- Challenges Selling RAG in Production: Members expressed frustration over the difficulty of convincing stakeholders about the value of RAG (Retrieval-Augmented Generation) in production environments.
- It's so hard to make ppl believe in that captures the ongoing struggle to gain traction for RAG implementations.
- Strategies for Document Updates: Managing frequent document updates raised challenges, which led to discussions on utilizing a vector database for automation.
- Suggestions included leveraging Qdrant for indexing and scheduling cron jobs to facilitate timely updates.
- LlamaDeploy & LlamaIndex Compatibility Confirmed: Members confirmed that LlamaDeploy is compatible with the latest version of LlamaIndex Workflow, ensuring seamless version syncing.
- They noted that deploying multiple workflows in LlamaDeploy effectively manages concurrent requests due to its asynchronous design.
Cohere Discord
-
Cohere Community is Coherent: Members lauded the Cohere community for its quality discussions, contrasting it with other AI communities that lack clarity.
- One member sought collaboration opportunities within this vibrant environment.
- Excitement for Cohere Research Innovations: The community is abuzz about recent advancements in Cohere research, with users reporting substantial progress.
- Developments are being quickly rolled out, marking a significant milestone for the team.
- Understanding Song Embedding Functionality: Inquiry arose regarding the Song Embedding Notebook, specifically on calculating recommendations with song IDs.
- Members discussed whether sentence2vector or word2vec was the method of choice for developing these embeddings.
- Diving into Aya vs Command Models: Discussions clarified that Aya is optimized for multilingual tasks, while Command focuses on production environments.
- Members noted that Aya excels specifically in multilingual capabilities, leading to a productive discussion.
- Fix the Weird JSON Argument Bug: A member raised concerns about JSON formatting errors in function calls, highlighting issues with single versus double quotes.
- Frustration mounted over this odd bug, as another member stressed the importance of proper JSON escaping with examples.
OpenInterpreter Discord
-
Open Interpreter fixes just dropped: Recent updates to
interpreter --osare now available on pip, inviting users to test for issues before launching voice mode.- The updates aim to enhance the user experience for those facing challenges with the interpreter.
- Rate limits frustrations with Claude: Members reported feeling hindered by Claude's rate limits, which are causing workflow interruptions.
- One member humorously pointed out that the rate limits are truly testing their patience.
- Setting up custom OpenAI API agent: There's an ongoing discussion about the feasibility of configuring a custom OpenAI API agent instead of using Claude.
- Documentation to assist users in setting up their configurations has been shared for practical guidance.
- Clevrr-Computer Empowers AI Productivity: Clevrr-Computer offers an open-source implementation of Anthropic's Computer for performing base tasks with AI agents.
- The project is celebrated for its potential to automate tasks and enhance productivity across various platforms.
- Explore Chrome's Built-in AI Features: A link to Chrome's Built-in AI resources showcases powerful integrations of AI within web activities.
- These features promise to improve user interaction with sophisticated AI tools directly embedded in the browser.
LAION Discord
-
Video Model Training Bottlenecks Observed: Users reported serious delays when training video classification models on 8 GPUs, primarily due to dataloading bottlenecks with 7M frames in MP4 files.
- Converting these files to JPEGs would dramatically expand the dataset size to 1TB, exacerbating performance issues.
- DataLoader Optimization Tips Shared: Community suggestions emphasize the importance of monitoring DataLoader performance by timing data fetches against GPU processing.
- Implementing effective prefetching strategies is vital for keeping up with faster GPU speeds, minimizing bottlenecks.
- Disk IO Discussions Affecting Training Speed: Concerns arose regarding whether SSD or HDD configurations lead to significant read speed or IOPS bottlenecks during training.
- Monitoring disk IO is crucial to diagnose potential issues impacting DataLoader performance and overall training efficiency.
- Importance of Model Size for Training Efficiency: Users discussed using a 50M parameter model that led to delays when working with larger batch sizes, indicating insufficient capacity for processing video data.
- It was suggested that increasing model size could alleviate data loading issues, enhancing overall performance.
- New Webinar on LLM Application Best Practices: A popular YouTube webinar titled Best Practices for Building Successful LLM Applications has gained nearly 1000 views in its first day, presented by a Senior ML Engineer from Meta.
- The session promises valuable insights on LLM implementation tailored for effective and impactful applications, encouraging hands-on learning.
OpenAccess AI Collective (axolotl) Discord
-
DPO Evaluations Made Simple: You can perform evaluations for Direct Preference Optimization (DPO) using the Axolotl codebase by loading datasets and comparing predictions to ground truth with the
load_prepare_dpo_datasetsfunction.- Efficiency meets accuracy; ensure your DPO model runs in evaluation mode with
model.eval()before generating predictions. - Generating Efficient Predictions: Utilize torch's no_grad context to generate predictions from the evaluation dataset, optimizing memory usage by not tracking gradients.
- This approach fosters memory-saving predictions, ensuring smooth and efficient evaluation processes.
- Metric Calculation with Ease: After generating predictions, calculate various metrics like accuracy or F1 score using scikit-learn, specifically through functions like
accuracy_score.
- This enables precise comparisons between predicted and true labels, reinforcing the evaluation integrity.
- Integrate Callbacks for Streamlined Training: Integrate evaluation into training using callbacks such as
BenchEvalCallback, allowing for evaluations at predefined intervals.
- This smooth incorporation of metrics helps maintain an efficient training routine, ensuring continuous monitoring of model performance.
- Efficiency meets accuracy; ensure your DPO model runs in evaluation mode with
Interconnects (Nathan Lambert) Discord
-
Polls on Mid Training Content Spark Discussion: Members initiated a discussion on mid training, questioning what precisely is included, which defined the scope and processes involved.
- Everything which is specialized training on some data but not RLHF, led to deeper exploration of methodologies.
- Epoch Specifics: Training on Coding: The conversation featured a suggestion that mid training could involve training for 1-2 epochs specifically on coding, clarifying distinctions within training methodologies.
- This aimed to enhance understanding of epoch training impacts on AI performance.
- Diversity in Historical Mails Discussed: A member noted that diversity should be injected into historical mails, indicating an interest in data variation and its implications.
- This calls into question how historical datasets inform current AI models.
- Memes Make Waves in AI: A member linked to a tweet, potentially highlighting cultural commentary within the AI community.
- Though specifics weren’t provided, memes often serve as a unique lens on technical discussions.
LangChain AI Discord
-
Evaluating Datasets from PDF Files: A member inquired about methods for evaluating and managing datasets, specifically targeting PDF data, as they intend to run evaluations with a PDF file.
- This poses challenges regarding the methodologies for structured evaluations with unstructured formats, prompting discussions on potential approaches.
- Job Opportunity for AI Wizards: A member is actively seeking a solid AI developer for upcoming projects, highlighting a need for skilled talent.
- This engagement led to questions about potential project ideas that could capitalize on such expertise, fostering a brainstorming environment.
LLM Agents (Berkeley MOOC) Discord
-
Timestamp Clarification on Submission Email: A member noted the timestamp of the form email as Sep 28, 6:50 PM PST, providing clarity on the email submission context.
- This detail arose in addressing a specific issue with email submissions, highlighting the importance of accuracy in timestamps.
- Progress on Email Confusion: Another member confirmed they found the email and expressed optimism about a resolution moving forward.
- Their positive outlook suggests that the confusion surrounding the email issues is on the brink of being resolved.
DSPy Discord
-
MIPROv2 Enhances Prompt Generation: A member shared a quick thread on 'automatic prompt generation' using techniques from MIPROv2 optimizer with the GSM8K dataset.
- The implementation includes three modules for demo generation, instruction creation, and final prompt compilation to streamline the process.
- Three Modules for Structured Prompt Creation: The program consists of Module 1 for demos, Module 2 for instructions, and Module 3 for synthesizing the final prompt.
- This modular approach focuses on efficiency in prompt generation, leveraging a systematic structure to improve overall effectiveness.
LLM Finetuning (Hamel + Dan) Discord
-
Edgar's Resource Check: Edgar expressed gratitude to c123ian for sharing useful resources related to LLM Finetuning that he plans to review.
- While specifics on the resources were not detailed, this exchange highlights the collaborative nature of the discussion in the channel.
- Collaboration on LLM Techniques: Members engaged in discussions about different techniques and methodologies for Finetuning LLMs, showcasing varied expertise.
- Contributions emphasized the need for sharing actionable resources to improve model performance.
Torchtune Discord
-
Torchtune GitHub Gets New Issue: A new issue concerning various enhancements and fixes has been reported on the Torchtune GitHub, highlighting the need for community contributions.
- Members are encouraged to participate in addressing these enhancements, although the issue isn't specifically labeled for community help.
- Call for Collaboration on Torchtune: Interest in Torchtune is growing as members express a desire to collaborate on the recent issue regarding enhancements and fixes.
- The ongoing discussion centers around how the community can support the project, fostering an engaging collaborative atmosphere.
Mozilla AI Discord
-
AI Creators Push for Compensation Rights: Creators across the internet confront a crisis where their work fuels AI systems without consent or compensation, highlighting the necessity for an effective licensing platform.
- This emerging system aims to enable individuals to license their content for AI training, promising improved fairness for content creators.
- Human Native AI Launches Data Marketplace: Co-founder James Smith announced that Human Native AI is developing a data marketplace where creators can pool their works and receive fair compensation for AI training.
- This initiative seeks to address the inequality in data usage and provide assurances to content creators concerned about the exploitation of their works.
- Mozilla's Data Futures Lab Speaker Series Event: The talk featuring James Smith is part of Mozilla's Data Futures Lab Speaker Series, aimed at discussing equitable data ecosystems in the AI landscape.
- Participants are encouraged to RSVP for this event to engage in critical discussions about the future of data and generative AI.
Gorilla LLM (Berkeley Function Calling) Discord
-
Gorilla LLM Function Calling Insight: A concise point was made about Gorilla LLM concerning its function calling capabilities, indicating a significant improvement during discussions on Berkeley Function Calling.
- Good catch highlights that the team is keenly analyzing the nuances of the latest updates, potentially leading to enhanced model interaction.
- Potential Enhancements Discussion: Engineers noted that the functionality of LLMs continues to evolve, with emphasis on improved function calls becoming a priority in upcoming releases.
- This could lead to further optimizations, and participants are eager to see practical outcomes from these discussions.
The Alignment Lab AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The MLOps @Chipro Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The DiscoResearch Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
The AI21 Labs (Jamba) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
PART 2: Detailed by-Channel summaries and links
The full channel by channel breakdowns have been truncated for email.
If you want the full breakdown, please visit the web version of this email: !
If you enjoyed AInews, please share with a friend! Thanks in advance!