[AINews] AI Discords Newsletter 11/28/2023
This is AI News! an MVP of a service that goes thru all AI discords/Twitters/reddits and summarizes what people are talking about, so that you can keep up without the fatigue. Signing up here opts you in to the real thing when we launch it 🔜
Latent Space Discord Summary
- Discussion about seeking recommended references but without any specified context.
- User
@swyxiohinted at something referred to as "CodeForward" without any further context. - Exploration for resources on the topic of context management in AI, particularly with a focus on mathematical modelling or theory in multi-agent shared memory systems, as introduced by
@slono. - Anticipation for an upcoming AI Engineering book from OReilly, suggested as a potentially useful resource by
@swyxioand@henriqueln7. - In the paper club,
@eugeneyanrequested@yikesawjeezto select this week's papers for pre-reading. The proposed topics were papers on prompt improvements (arxiv.org/abs/2311.09277) and (arxiv.org/abs/2311.05997), which later changed to a RAG (Retrieval-Augmented Generation) context management paper (arxiv.org/abs/2311.09210) after a reminder about needing long-context methods. - Expressions of interest in discussing Lookahead Decoding in the future by
@slonoand a positive response to the current topic selection by@swizec. @semra5446suggested an additional future paper for discussion: arxiv.org/pdf/2311.01906.pdf.- Creation of a LangChain template by Harrison was mentioned for the paper Skeleton-of-Thought, shared by
@yikesawjeez.
Latent Space Channel Summaries
▷ Channel: ai-general-chat (8 messages):
- Recommended References: User
@coffeebean6887initiated a discussion about recommended references, although no specific context was given. - CodeForward: User
@swyxiomentioned something called "CodeForward", but didn't specify what it is or its relevance. - Context Management in AI: User
@slonois looking for resources related to the mathematical modelling or theory behind context management in AI, specifically in multi-agent shared memory systems. - OReilly's AI Engineering Book: Both
@swyxioand@henriqueln7hinted at an upcoming book on AI Engineering from OReilly. Their excitement may be due to expected high-quality content or relevance to the discussion topics.
▷ Channel: llm-paper-club (11 messages):
- Paper Selection for Discussion:
@eugeneyanasked@yikesawjeezfor the paper selection for the week to allow members time to pre-read.@yikesawjeezinitially proposed papers on prompt improvements (arxiv.org/abs/2311.09277) or (arxiv.org/abs/2311.05997). After@eugeneyan's reminder about studying long-context methods,@yikesawjeezproposed a RAG (Retrieval-Augmented Generation) context management paper instead (arxiv.org/abs/2311.09210). - Interest in Future Discussion Topics:
@slonoexpressed interest in discussing Lookahead Decoding (lmsys.org/blog/2023-11-21-lookahead-decoding/) in the future.@swizecexpressed enthusiasm for the topic selection in the paper club this week. - Additional Paper Suggestion:
@semra5446suggested another interesting future paper for discussion: arxiv.org/pdf/2311.01906.pdf. - LangChain Template:
@yikesawjeezalso mentioned a LangChain template made by Harrison for the paper Skeleton-of-Thought.
OpenAI Discord Summary
- Users across channels reported instances of performance issues, erratic behaviors, and possible limitation of GPT-4, triggering speculation about its functionality, early stage glitches, and server capacities. User
ahlattshared a specific instance with GPT-4 prematurely stopping aTradingEnvPython class generation. - AI and Language bias and limitations discussion initiated by user
posina.venkata.rayudu, referencing a Nature Human Behaviour article..doozemphasized that language limits in LLMs arise from technical aspects rather than a plan to enforce hegemony. - Use of ChatGPT for varied applications, from scenario modeling and exercise design to producing content for Dungeons and Dragons games and academic papers.
- Several users experienced problems with OpenAI's service login and performance, leading to speculations ranging from high server load to scheduled maintenance and updates.
- Several users were found expressing interest in GPT-4 and GPT-5, sparking light-hearted speculations about a possible GPT-5 during server downtimes.
- GPT-4 Store delay was discussed by various users including
.xiaoayi,dkracingfan,kyleschullerdev_51255, andlodosdwith concerns about the ambiguous timeline of the launch. - Possibility of Cooperation between OpenAI's large models and Boston Dynamics' robots was speculated by
@davidvon, with extended discussions on the level of consciousness and self-awareness in AI and AGI, specifically GPT models. - Users highlighted issues with OpenAI services, such as issues with image generation tools, the inability to find plugin options, lost conversation history, and specific DALL.E experiments with errors.
- Discussions revolved around effective prompt engineering and plugin usage, with instances such as user
komal0887inquiring how to produce non-evaluative sentences using the "gpt-3.5-turbo-instruct" model andleoalvarengaseeking plugin recommendations. NastyTimexpressed significant dissatisfaction with GPT-4's service and customer support, with other users suggesting alternative options and advising contacting OpenAI's customer support.- @xsrusapa and @solbus held a brief discussion regarding advice on integrating Slack with custom GPTs in order to ask questions and receive answers back via Slack.
OpenAI Channel Summaries
▷ Channel: ai-discussions (101 messages):
- GPT-4 Performance Discussion: Users
@3hx,@cr7462,@xyza1594,@JohnPringle, and@dogdroidexpressed concerns about apparent performance issues and possible "nerfing" of GPT-4 generation. They discussed instances where the model required prompt refinement, stopped code generation prematurely, and other erratic behaviors.@ahlattshared a specific example where GPT-4 prematurely stopped generating aTradingEnvPython class. - AI and Language Discussion: User
@posina.venkata.rayudubrought up a recent article from Nature Human Behaviour about AI and language problems, indicating Chatgpt handles approximately 100 languages but expressing frustration over article's publishing tactics. Another user,.dooz, responded that language bias in LLMs is due to technical limitations rather than a secret plot to hegemonize. - Discussion on Maxi's Introduction: User
@maxiisboredintroduced themselves and interacted with@posina.venkata.rayuduand@jclove_. - Customer Dissatisfaction:
@NastyTimexpressed significant dissatisfaction with the GPT-4 service and frustrations about perceived lack of customer support from OpenAI. Responders.dooz,@kesku, and@satanhashtagrecommended other options (GPT-4 API,BingChat, prompt engineering techniques), clarified the beta status of the product, and suggested to contactsupport@openai.com. - AI and AI Data Analyst Products: User
@tonyfancherinquired about AI data analyst products. User@fran9000shared a link to Sean Carroll's podcast discussing the current state of LLMs and their future.@jaimdreported that Bing chat appeared to be filtering conversations about intelligent AIs.
▷ Channel: openai-chatter (459 messages🔥):
- Use of ChatGPT for Diverse Applications: Users discussed the diverse ways they use ChatGPT, from @aloy2867 looking for therm scenarios and exercises, to @satanhashtag using it for their DnD game, to @xyza1594 suggesting it could be used as a tool for writing papers.
- ChatGPT performance and login issues: Beginning with @cjdaigle2 experiencing login issues, many users reported ChatGPT being slow, down, or throwing errors. Some speculated that the problem was due to high server load or momentary downtime for maintenance and updates.
- Issues with data and features: Users like @winksaville, @pruo, and @thefoodmaster raised issues, such as problems with image generation tools, ChatGPT skipping some answers, the inability to find plugin options, and lost conversation history.
- Interest and speculation about upgraded versions: Users like @m_12091 and @chadgpt showed interest in upgraded versions like GPT-4 and GPT-5. Some users, like @chotes and @deltamza, humorously speculated about the possible release of GPT-5 during the downtime.
- Warning about spamming and content misuse: @kesku warned users about misrepresenting AI generated output as human generated in line with OpenAI's terms of service. Users @zinthaniel and @aryann warned about the possible temporary unavailability of some features and services due to high demand and the inadvisability of abusing refresh functions.
▷ Channel: openai-questions (206 messages🔥):
- Issue with Subscription and Plan Access: User
@alexgoganmentioned about losing his Plus membership and being moved automatically to the free plan. - Concerns About GPT Role Performing: User
@creriganhad a question on whether creating multiple GPTs with single jobs is better than having a single GPT to do related jobs.@solbussuggested trying both approaches and observing the outcomes. - Issues with Creating OpenAI Accounts: User
@phamngocanh9xreported inability to create an OpenAI account from Vietnam. - Web Scraping Suggestions Requested: User
@viprazasked for ways to scrape the web for articles, and included other related details such as article title, description and author's information. - Integrated Custom GPTs Inquiry: User
@xyrusapaasked for advice on using Slack to ask questions to a custom GPT and receive answers back into Slack. - Difficulty with Model: User
@dotailsstruggled with an issue!Unterminated string in JSON at position 548 (line 1 column 549).
▷ Channel: gpt-4-discussions (83 messages):
-
Training GPTs: Users such as
@haribala,@DawidM,@pietman, and@jaynammodidiscussed the limitations of training GPTs. It appears that the models cannot be "trained" in a traditional sense - rather, they rely on connecting to an external API, feeding instructions, and a knowledge base of files to customize responses. -
Launch of OpenAI GPT Store: Users
@.xiaoayi,@dkracingfan,@kyleschullerdev_51255, and@lodosdexpressed concerns about the delayed launch of the official OpenAI GPT store. As of the end of November, no official news or updates relating to the launch timeline had been released. -
API Integration and Enterprise Workflow:
@phild66740,@solbus, and@kyleschullerdev_51255discussed possibilities to integrate GPTs into an enterprise workflow, particularly re-purposing custom GPTs for API use. -
Chat Functionality Issues: Towards the end of the discussion, users like
@halalarax,@rangerslayer97,@hipo_., and others reported temporary outages and issues with ChatGPT including unexpected responses. -
Integrating plugins with Assistant API: Lastly, user
@zaprime.inquired about using plugins with the assistant API, but no response was given at the time of transcript cutoff.
▷ Channel: prompt-engineering (30 messages):
-
Plugins Recommendations: There is a request by
@leoalvarengaasking for plugin recommendations, but no recommendations were provided. -
Prompt Engineering:
@komal0887asked for assistance in writing a prompt that would generate articles without evaluative sentences using gpt-3.5-turbo-instruct model. -
Discussion on OpenAI and Robotics: A conversation initiated by
@davidvonspeculated on the possibility of OpenAI's large models cooperating with robots, potentially giving them physical bodies. -
Debate on AGI and Self-Awareness: There was a heated discussion led primarily by
@jaynammodi,@.pythagoras, and@syndicate47debating the extent of consciousness and self-awareness in AI and AGI, specifically GPT models. -
Issues with DALL.E:
@kh.drogonlamented difficulties in getting DALL.E to correctly spell words on an image. -
Extracting Information from Transcripts:
@greenysmacasked for guidance on refining a prompt to correctly identify the speaker count from a transcript. Suggestions from@notsoluckyduckyand@solbusproposed the use of advanced data analysis.
▷ Channel: api-discussions (30 messages):
- Plugin Recommendations:
@leoalvarengaasked for plugin recommendations. No further details or responses were given. - Generating Non-Evaluative Articles:
@komal0887sought help with writing a prompt to generate an article without evaluative sentences using the "gpt-3.5-turbo-instruct" model. - Cooperation with Boston Dynamics' Robots: In a series of comments,
@davidvonspeculated on the potential cooperation between OpenAI's large models and Boston Dynamics' robots to allow AI to have a physical body. - DALL.E's Spelling Issues:
@kh.drogonshared an issue with DALL.E misspelling words when creating logos. The user sought advice, but no responses were present. - OpenAI Model's Consciousness Debate: A debate occurred between
@davidvon,@jaynammodiand@.pythagorasregarding whether OpenAI's large models have free will and self-awareness, revolving around a GPT model-operated robot.@eskcantaand@syndicate47offered critical perspectives on these assertions. - Counting Speakers in a Transcript:
@greenysmacasked for help in refining a prompt that counts speakers in a transcript.@notsoluckyduckyrecommended using advanced data analysis to parse the names, remove duplicates, and return the list length. An additional validation was provided by@solbus.
LangChain AI Discord Summary
- Discussions about using and configuring LangChain for various functionalities and fine-tuning procedures:
- Concern about the user interface of LangChain agents shared by
@moooot, showing interest in more intuitive UIs for displaying information. - Request for instructions on connecting a LangChain agent to Zapier, posed by
@jonanz. @steve675questioned the possibility of creating a Q&A assistant for tabular data in LangChain.- Technical issues faced by
@menny9762with passing a prompt into a ConversationalRetrievalQAChain and setting the{lang}variable, with@seththunderproviding input. @l0st__sought assistance on fine-tuning a LangChain agent that access SQL databases, and discussions about "stuck" loops and delayed responses.@sampson7786asked how to request an invite code for LangSmith, with suggestions given by@seththunder.
- Concern about the user interface of LangChain agents shared by
- Tutorials, resources, and work sharing related to LangChain:
- A new tutorial shared by
@kulaoneabout deploying LangChain on Cloud Functions using Vertex AI models, available on YouTube. @joshuasundancepromoting the latest version of BERTopic that now supports LCEL runnables, with link provided to GitHub Release.@andysingaldiscussing the role of LangChain+LlamaIndex in extracting tables and images using Multimodal LLM, citing a Medium article as reference.- Additional request from
@l0st__for guidance on fine-tuning a LangChain agent, providing an example conversation flow.
- A new tutorial shared by
LangChain AI Channel Summaries
▷ Channel: general (52 messages):
- Request for Langsmith Invite code: @sampson7786 asked about how to request for an invite code for Langsmith.
@seththundersuggested that it's done from the LangChain website. - Langchain Conversational Retrieval QA Chain: @menny9762 discussed issues with passing a prompt into a
ConversationalRetrievalQAChainand figuring out how to set the variable{lang}.@seththundersuggested that anything in{}needs to be set as input variables. - SQLDatabase Agent: @l0st__ shared the need for assistance on fine-tuning a LangChain agent that has access to a SQL Database to act as a sales assistant, mentioning about issues of getting "stuck" in loops and delayed responses.
@gloria.mart.lurecommended SQLDatabase agent & toolkit for usage. - User Interface for LangChain: @moooot expressed concern on why the user interface for agents is essentially a chatbot and inquired about more intuitive UIs for displaying the information fetched by agents.
- Integration of LangChain with Zapier: @jonanz asked for guidance on how to connect a LangChain agent to Zapier, referencing the deprecated link on the LangChain website. A solution was not provided in the chat.
▷ Channel: langchain-templates (2 messages):
- Creating Q&A Assistant for Tabular Data: User
@steve675inquires if LangChain allows for creating a Q&A assistant specifically tailored to tabular data. His need is for a chatbot which can interact with multiple tables, given that each table may have varying columns.
▷ Channel: share-your-work (5 messages):
- BERTopic Runnable Support:
@joshuasundanceshared the latest version of BERTopic which now supports LCEL runnables, highlighting its usefulness and expressing excitement for its potential applications. - LangChain Agent Fine-Tuning Request:
@l0st__sought assistance for fine-tuning a LangChain agent that can act as a sales assistant and access SQL databases, providing a detailed use-case as an example. - Multimodal LLM for Data Extraction:
@andysingalshared a Medium article discussing the role of LangChain+LlamaIndex in extracting tables and images using Multimodal LLM.
▷ Channel: tutorials (2 messages):
- LangChain Deployment Tutorial: User
@kulaoneshared a new tutorial on how to deploy Langchain on Cloud Functions using Vertex AI models for scalability. The tutorial can be watched on YouTube. - Fine-Tuning LangChain Agent: User
@l0st__expressed a need for assistance in fine tuning a LangChain agent that has access to a SQL Database and would act as a sales assistant. He provided an example conversation flow for clarification.
Nous Research AI Discord Summary
- Discussion surrounding the upload and processing of the MJ_dataset on Hugging Face by
@tsunemotowho ran into upload speed limitations. This led to a conversation on rate limiting and possible ways around it with the new Hugging Face account. - Random Twitter and YouTube links shared by
@.andrewncand@pradeep1148without any contextual details. - Conversations on tinygrad development and the YouTube stream of George Hotz building upon openhermes, shared by
@yorth_night. - Introduction of the new compute provider, Voltage Park, by
@coffeebean6887, with its connections to Imbue, Character, Atomic, and an unannounced $500m funding round. - Variety of Twitter posts, and research links shared by
@papr_airplane,@if_a, and@oozyi. - Extensive conversations on performance and fine-tuning of chatbot models in the context of FFT, QLORA, DeepSpeed, Axolotl, FSDP and the difficulties of training large models.
- User experiences of GPT-4 highlighting its lackluster performance and the need for more user input.
- Updates on recent model tunes and testing, specifically the "koishi-120b-qlora" by
@ludis___. - Technical challenge and subsequent discussion on running inference for StyleTTS2 shared by
@nemoiaand@lhl. - A focus on the limitations of model benchmarking and potential manipulation of results.
- Discussions on building high-PC specs for LLMs with focuses on CPU, RAM, and PCIe lane considerations.
- Queries on open-source models optimized for planning and function call abilities, however, no specific model was recommended.
- Comparisons of sentence similarities recommended by
@philpaxusing sentence embedding and cosine similarity operation, with a link shared to Sentence-BERT's documentation. - In-depth discussion on the performance of various models on different GPUs and approaches for pruning and quanting including ExlLlama and GGUF, with an emphasis on ExlLlama working better with 2 GPUs for large models.
- Shared link to a quanted model of Hermes 2 with exllama-2: https://huggingface.co/bartowski/OpenHermes-2.5-Mistral-7B-exl2
Nous Research AI Channel Summaries
▷ Channel: off-topic (7 messages):
- Sharing of MJ_dataset:
@tsunemotoshared a link to his MJ_dataset on Hugging Face, which contains a million rows from a midjourney dataset. It took him approximately 3 days to process and upload due to restrictions on upload speeds. - Discussion on upload speed limitations:
@tsunemotoand@crainmakerdiscussed the issues of rate limiting on Hugging Face.@tsunemotomentioned that despite trying to use the Rust-based library for transferring data to Hugging Face by setting the hf transfer environment variable, he still experienced limitations possible due to his new Hugging Face account. - Links shared:
@.andrewncshared a Twitter link without any additional context.@pradeep1148shared a YouTube link also without any additional context.
▷ Channel: interesting-links (10 messages):
- Tinygrad Progress:
@yorth_nightcommented that tinygrad is looking more and more sexy with@vatsadevagreeing, getting better every day. - George Hotz Stream:
@yorth_nightshared a YouTube link of George Hotz working with openhermes as a base for "Q*". - New Compute Provider Available:
@coffeebean6887introduced a new compute provider, Voltage Park, stating it received a large part of its funding from Imbue and also mentioning contracts with Character and Atomic. Additionally, Voltage Park is reported to provide Burst and On-Demand h100s. - Voltage Park Funding:
@coffeebean6887also shared that the new compute provider, Voltage Park, had an unannounced $500m funding round. - Shared Research and Social Media Links: Links were shared by
@papr_airplaneand@if_a(Twitter posts, another Twitter post) and@oozyi(Research paper).
▷ Channel: general (232 messages🔥):
-
Performance and Finetuning of Chatbot Models: Users explored model performance and finetuning methods, with discussions revolving around FFT, QLORA, DeepSpeed, Axolotl, FSDP and regarding difficulties of training large models and running out of GPU memory.
@tekniumshared one of his wandb runs for a QLORA over Hermes 2 experiment with the Trismegistus dataset [Link]. -
User Opinions on GPT-4: The community expressed some disappointment with GPT-4's behaviour, particularly noting it leaves more work for the user and appears to be providing shorter, incomplete generation results.
-
Model Recommending and Testing:
@ludis___announced completion of a 120B tune on instruct data named "koishi-120b-qlora", and the community showed interest in model Deepseek, particularly the 7B version.@vatsadevremarked on OpenHermes becoming a strong brand in local models, recognized by half of Twitter as the go-to replacement for ChatGPT. -
StyleTTS2 Troubleshooting:
@nemoiasought advice on running inference for StyleTTS2.@lhlshared his notes on StyleTTS2 setup for reference [Link]. -
Model Benchmarks: The community discussed the use and potential shortfalls of benchmarking as a measure of model performance, raising issues such as overfitting to benchmarks and manipulation of benchmark results.
▷ Channel: ask-about-llms (66 messages):
-
PC Specs for LLMs:
@orabazesinitiated a discussion about building a high spec PC for LLMs and asked whether using a Threadripper CPU is necessary. Various users weighed in, with@night_w0lfsuggesting that any latest generation CPU should suffice and@.andrewncrecommending speccing out a Lambda Labs workstation. The importance of having enough system RAM and the right number of cores were also emphasized. -
LLM Planning and Function Calling:
@htahskaasked for recommendations on open source models that are optimized for planning and function calling abilities. No model was specifically recommended in the available excerpts. -
Graphics Cards and PCIe Lanes:
@philpaxprovided key information about the limitations of x670 having only 24 PCIe lanes and thus impacting the number of GPUs one can fit. -
Autonomous Browser-Based LLM Project:
@htahskaasked for suggestions about LLMs, focusing on their planning and function calling abilities. No specific suggestions were made in the responses. -
Comparing Sentence Similarities:
@philpaxrecommended@asada.shinon, who wanted to calculate the distance between sentences, to use sentence embedding and perform a cosine similarity operation. A link to Sentence-BERT's documentation was shared: https://www.sbert.net/docs/quickstart.html#comparing-sentence-similarities. -
GPU Performance and Quanting Tools: There was an extensive discussion about the performance of various models on different GPUs, notably with
@tekniumand@coffeebean6887. Different approaches for pruning and quanting, including ExlLlama and GGUF, were discussed; it was noted that ExlLlama was working better with 2 GPUs for large models. -
Link Shared: Font to a model of Hermes 2 quanted with exllama-2: https://huggingface.co/bartowski/OpenHermes-2.5-Mistral-7B-exl2
NB: LLMs = Large Language Models, ExlLlama and GGUF = Quanting tools, GPUs = Graphics Processing Units, PCIe = Peripheral Component Interconnect express.
Alignment Lab AI Discord Summary
- A message from user
@bryhen22on #looking-for-collabs seeking collaboration without providing further context. - User
@magusartstudiosand@ldjdiscussed a specific YouTube video shared in #general-chat. The content of the video was described as being cursed. - The same YouTube video link was shared by user
@magusartstudiosin #fasteval-dev without additional commentary or context.
Alignment Lab AI Channel Summaries
▷ Channel: looking-for-collabs (1 messages):
- User
@bryhen22tried to send an unidentified recipient a message. Further context or detail is not provided in the message.
▷ Channel: general-chat (2 messages):
- User
@magusartstudiosshared a YouTube link. - User
@ldjdescribed the content as cursed.
▷ Channel: fasteval-dev (1 messages):
- User
@magusartstudiosshared a YouTube link: https://www.youtube.com/watch?v=rMBLZtPmlsQ
Skunkworks AI Discord Summary
Only 1 channel had activity, so no need to summarize...
Skunkworks AI Channel Summaries
▷ Channel: off-topic (1 messages):
@pradeep1148shared a YouTube video link.
LLM Perf Enthusiasts AI Discord Summary
-
Active discussions on GPT-4's code generation capability and inference speeds. Concerns include incomplete function implementations and deteriorating speeds (particularly during peak hours), observed by
@potrock,@pantsforbirdsand@res6969. A hypothesis by@evanwechslerand@nosa_.posits potential cost and latency optimizations compromising model response. API performance above the cursor still holds as per@res6969. There was shared agreement on the increased necessity for prompt engineering to maximize performance, with@thebaghdaddysuggesting a resource by Lilian Weng on the topic. -
In the Open Source Models section,
@thisisnotawillsought recommendations for storytelling models, with@pantsforbirdssuggesting Mistral or its variants, also sharing their success with a custom retrieval model for their RAG pipeline. Links to fine-tuned versions of Mistral 7B like Dolphin-2.1 and new models like Starling-LM-7B-alpha were shared, along with a Twitter thread offering additional context. -
Tangential observations on the credibility of AI rumors and image generation patterns under scope by
@pantsforbirdsand@justahvee. Notably, the 'more fantasy' specification leading to a psychedelic space vibe in images was sought to be analyzed. -
Users
@thisisnotawill,@justahvee, and@pantsforbirdswere discussing multiple aspects of an AI Dungeon Master in the collaboration channel. Challenges identified include the difficulty of maintaining a cohesive storyline over hundreds of messages and balancing player freedom with plot coherence. Simplifying the problem-solving approach and adjusting current methods were proposed as potential solutions. -
@jeffreyw128discussed the role of traditional search providers and the significance of vector and keyword search for speed in their project "metaphor". -
Creation of a prompt channel was proposed by
@pantsforbirdsin the feedback-meta channel. -
Detailed discussion on the cost structure of running a PDF-to-RAG/LLM pipeline was led by
@res6969, who revealed splits between Azure OCR and GPT-4 and a drift towards a growing share of OCR costs. Transition from AWS Textract to Azure OCR for table extraction capability was also shared.@pantsforbirdsrevealed challenges with Nougat with a switch to AWS Textract mulled, further sharing a link discussing Amazon Textract's new layout feature. -
A user inquired about others located in Vermont.
-
Issues about OpenAI's speed and inability to login were raised by
@robhaisfieldand@iyevenko. Desire for a 70b model from Mistral was expressed to reduce dependency on OpenAI. An incident disrupting services at OpenAI was shared by@pantsforbirds, linking to the overall status and the specific incident.
LLM Perf Enthusiasts AI Channel Summaries
▷ Channel: gpt4 (28 messages):
-
Code Quality in GPT-4:
@potrockraised concerns about code generation quality in GPT-4, stating it has become "hilariously bad" and often leaves placeholder comments or refuses to implement functions completely. This was echoed by@pantsforbirdswho shared the frustration of GPT-4 only completing part tasks. -
Inference Speed Issues: Both
@pantsforbirdsand@res6969noted worsening inference speeds, particularly during peak business hours. -
Possible Optimization for Cost & Latency: Users like
@evanwechslerand@nosa_.speculated that OpenAI may be optimizing GPT-4 for cost and latency, possibly compromising on response length or quality. -
Consistent API Performance: According to
@res6969, API performance, particularly for applications above the cursor, has so far remained unaffected. -
Prompt Engineering:
@pantsforbirdsand@thebaghdaddyagreed that with the latest models, fine-tuning the prompt has more of an impact on performance than typical engineering.@thebaghdaddyshared a link to a post on prompt engineering by Lilian Weng for further reading.
▷ Channel: opensource (7 messages):
- Open Source Models for Creative Writing/Storytelling:
@thisisnotawillinquired about open-source model recommendations for creative writing and storytelling.@pantsforbirdssuggested Mistral or its fine-tuned variants, specifically citing their success with a custom retrieval model for their RAG pipeline using Mistral 7b, fine-tuned on SageMaker. - Fine-tuning Models:
@pantsforbirdsshared their experience of fine-tuning Mistral 7B themselves using SageMaker and suggested that there are open-source tunes of Mistral 7b available which can serve as a good starting point for broader use-cases. - Resource Links:
@pantsforbirdsshared a link to a fine-tuned version of Mistral 7B, namely Dolphin-2.1, on Hugging Face, stating that it performed well in their experience. - Additional Model Recommendation:
@pantsforbirdsalso mentioned a new model, Starling-LM-7B-alpha, on Hugging Face. They also shared a Twitter thread, providing additional context, but clarified that they had not personally tested the performance of this model.
▷ Channel: offtopic (4 messages):
- Credibility of AI Rumors:
@pantsforbirdsand@justahveediscussed the validity of some rumors about AI. While@pantsforbirdsdeclared an overall skepticism,@justahveefound an email allegation more plausible than a 'random post', but still questioned the feasibility of skipping towards a high-level test like exploiting AES. - Image Generation Patterns:
@pantsforbirdsbrought up a pattern noticed in image generation where continual requests for enhancing a certain aspect (e.g., 'more fantasy') leads to a result resembling a psychedelic space image, and was seeking input on the phenomenon.
▷ Channel: collaboration (11 messages):
- AI Dungeon Master Building Challenges: User
@thisisnotawillshared their experience with developing an AI Dungeon Master, revealing the difficulty of maintaining a cohesive storyline over hundreds of messages. Despite using hierarchical summarization and RAG, the AI seems to falter in plot continuity. Concerns include the AI repeatedly referring back to past events or veering into an unexpected direction. - Balancing Player Freedom and Plot Continuity:
@thisisnotawillexpressed the need for an AI solution that can maintain the narrative's coherence while still allowing for player freedom and a dynamic plotline. The challenge is managing the fine balance between AI-guided plot arcs and player choices. - Simplifying Problem Solving Approach:
@justahveesuggested narrowing the focus to solve the most crucial problems first. They advised isolating a single story arc and successfully execute it end-to-end before delving into dynamic plots. The emphasis is on creating a system that works well with a particular set of user needs before expanding to a wider range. - Real-Time AI-assisted DMing: User
@pantsforbirdsrevealed their work on creating a real-time AI-assisted Dungeons and Dragons tool. The model in development aims to parse player conversations and furnish possible details on player queries. - Collaboration and Sharing: Both
@thisisnotawilland@pantsforbirdsexpressed interest in each other's projects and the potential for collaboration or learning from each other's experiences.
▷ Channel: speed (2 messages):
- Vector and Keyword search:
@jeffreyw128discussed the use of traditional search providers for keyword search in their project named "metaphor", highlighting its speed. They mentioned that they also have a neural, vector-based search system, and their project handles the return of HTML contents.
▷ Channel: feedback-meta (1 messages):
- Proposed Prompt Channel:
@pantsforbirdssuggested the creation of a prompt channel for discussion and sharing of prompts, indicating the potential benefits and utility of having such a centralized space.
▷ Channel: cost (15 messages):
- Cost Structure of Running a PDF-to-RAG/LLM Pipeline: User
@res6969shared an insight that while operating a PDF-to-RAG/LLM pipeline, OCR through Azure accounts for 52% and GPT-4's new pricing stands for 48% of total costs per document. It was also mentioned that the rate of OCR costs will increase over time while GPT-4's share will decrease. - Azure OCR vs Open Source Models: When asked about the accuracy of Azure OCR,
@res6969stated that it was essentially 100% accurate with the exception of occasional I's and L's. Open source models were said to be nearly as accurate but as expensive to run. - Migration from AWS Textract to Azure OCR:
@res6969shared that they moved from AWS Textract to Azure OCR due to the equivalent cost. Azure OCr was preferred as it provided higher rate limits on the beginning plan and better support for table extraction from documents. - Problems with Nougat: User
@pantsforbirdsraised concerns about facing production issues with Nougat, stating that it was particularly challenging to debug.@pantsforbirdsexpressed interest in testing AWS Textract due to its multicolumn support. - Experience with AWS Textract:
@pantsforbirdsshared a link to an AWS blogpost titled "Amazon Textract's new layout feature introduces efficiencies in general-purpose and generative AI document processing tasks".
▷ Channel: irl (1 messages):
- An user going by the handle
@thebaghdaddyasked if anyone in the channel is located in Vermont.
▷ Channel: openai (8 messages):
- OpenAI Platform Issues: User
@robhaisfieldraised an issue about OpenAI being slow.@iyevenkoconfirmed the problem and mentioned inability to log into the platform. - Wish for Mistral 70b Model:
@robhaisfieldexpressed a desire for a 70b model from Mistral to reduce dependence on OpenAI. - OpenAI Incident Information:
@pantsforbirdsinformed about an incident at OpenAI, linking this issue to some db problems on OpenAI's end. They provided links to the overall status and the specific incident for further details.
MLOps @Chipro Discord Summary
Only 1 channel had activity, so no need to summarize...
MLOps @Chipro Channel Summaries
▷ Channel: general-ml (3 messages):
- Cross-posting Content: User
@wangx123shared a YouTube video link across multiple channels. This led to a brief discussion among other users including@c.s.aleand@huikangabout the feasibility and need of a Discord rule to automatically remove post if posted onto multiple channels in a short period of time.
The Ontocord (MDEL discord) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
AI Engineer Foundation Discord Summary
- Discussed the importance of proper commit practices and code quality in the
#generalchannel.@kasparpetersonexpressed concern over direct commits to master, CI failures, andprettierandeslinterrors. Suggestions to improve these practices included adhering to Git PR flow, enforcing guidelines, and adding Github Actions for public repos to ensure quality.- "We had some straight commits to master, and some CI failures..." -
@kasparpeterson
- "We had some straight commits to master, and some CI failures..." -
- In the
#generalchannel,@._zacknowledged these issues, suggesting that they were a result of the current setup and invited a revision of the contributing guidelines. - Meeting participations and proposals were addressed by
@pwuts, showing willingness to answer queries regarding a previously proposed idea and intent to participate in upcoming meetings. - In the
#eventschannel, posts from@hackgooferand@._zincluded notifications and reminders of the AIEF Weekly Meeting and the AIEF/Agent Protocol meeting. - The topic of including authentication in the core protocol emerged as a discussion point in the
#eventschannel, with@kasparpetersonproposing its inclusion and sharing a detailed comment on GitHub. Members'@hackgoofer','@ntindle', and'@juanreds'displayed unified approval for integrating authentication into the core protocol.- "I recommend that authentication should be part of the core protocol..." -
@kasparpeterson
- "I recommend that authentication should be part of the core protocol..." -
AI Engineer Foundation Channel Summaries
▷ Channel: general (10 messages):
- Discussion on Commit Practices: In the conversation between
@kasparpetersonand@._z, the proper commit practices and use of precommit hooks were analyzed.@kasparpetersonnoted there have been straight commits to master and CI failures, along withprettierandeslinterrors, suggesting that the team should follow git PR flow.@._zexplained these errors might have occurred due to their current setup and suggested revising the contributing guidelines. - Efforts Toward Code Quality:
@kasparpetersonfurther proposed enforcement of existing guidelines rather than revising them. He suggested adding Github Actions for public repos to ensure quality, including a separate CI job for every push to validate the schema, using a simple example server and testing command. The objective would be to provide a firm safety to the core repository. - Planning for Future Meetings:
@pwutsexpressed his willingness to answer any queries about an idea he proposed, which was discussed in a previous team meeting. He also plans to attend the coming meeting.
▷ Channel: events (10 messages):
- Tomorrow's AIEF Weekly Meeting Announcement: User
@hackgoofershared a Discord invite link for an upcoming AIEF Weekly Meeting event. - Reminder of AIEF/Agent Protocol Meeting: User
@._zsent out a reminder about the AIEF/Agent Protocol meeting happening four hours after the message was posted. - Discussions on Including Authentication in the Core Protocol:
@kasparpetersonrecommended that authentication should be part of the core protocol and shared a link to his detailed explanation on GitHub. He indicated that it is optional to pass tokens in the Authorization header field. He also suggested to approach the implementation lean and skip the plugins and finish the info endpoint as the config options are optional anyway. Their comment can be found here. - User Responses to the Topic of Authentication: Other users
'@hackgoofer','@ntindle'and'@juanreds'shared their views on the topic. There was a general consensus in favor of including authentication in the core protocol, with'@ntindle'and'@juanreds'specifically stating they agree it should be part of the core protocol. - AIEF/Agent Protocol Meeting Starting Reminder: User
@._zposted a reminder message that the meeting was starting and prompted attendees to check the meeting notes and add any topics for discussion.
The Perplexity AI Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
YAIG (a16z Infra) Discord Summary
Only 1 channel had activity, so no need to summarize...
YAIG (a16z Infra) Channel Summaries
▷ Channel: ai-ml (2 messages):
- AI Research Video Discussion:
@nickw80recommended a YouTube video about the latest AI research. He highlighted that despite the title, the video covers recent studies and is worth watching.