[AINews] AI Discords Newsletter 11/21/2023
This is AI News! an MVP of a service that goes thru all AI discords/Twitters/reddits and summarizes what people are talking about, so that you can keep up without the fatigue. Signing up here opts you in to the real thing when we launch it 🔜
Latent Space Discord Summary
- Controversy over a show's promotional ad, with
@vcarlcriticizing its misuse of the term 'hard fork', highlighting a common misunderstanding between open source software forks and blockchain forks. - Consideration and exploration of alternatives to the OpenAI API due to concerns raised by
@RyanAB. Suggestions included Claude, Google, and services provided by Replicate / Hugging Face. Anyscale was also recommended due to their efficient and cost-effective services for open-source models (Anyscale link). - Announcement of Claude 2.1's release, with notable upgrades such as a 200k context, reduced hallucination rates, and the inclusion of a Tool use API as shared by
@tiagoefreitas. - Appreciation for Stable Video Diffusion by Stability.AI, a video shared by
@chris946097. While considering a possible demo bias, the user acknowledges the impressive quality of the video. - Report of GPT API issues by
@dsquared70and@vcarl, suggested to be a database problem by@slono; reference to reported issue on OpenAI's status page. The downtime led to@slonotesting Anyscale models and finding them useful. - Inquiry about the relevance of a GPT-3.5-turbo related link to a genie question posted by semianalysis, shared by
@guardiangfor conversation. - Sharing of a functional link to a one-click installer for machine learning model testing, despite slow performance on specific hardware, as pointed out by
@growthwtf. @growthwtf's suggestion that Controlnet offers better speed for working with machine learning models.
Latent Space Channel Summaries
▷ Channel: ai-general-chat (26 messages):
- Promotional Ad Misinterpretation:
@vcarlvoiced his dissatisfaction about a show's promotional ad, stating that it gives a "totally inaccurate definition of 'hard fork'", confusing open source software forks with blockchain forks. - Considerations of Alternatives to OpenAI API: Amidst recent drama surrounding OpenAI,
@RyanABaired his worries about relying on the OpenAI API. Looking for convenient and qualitative alternatives, he considered Claude and services provided by Replicate / Hugging Face.@coffeebean6887suggested considering Google, and highlighted Anyscale's infra costs and speeds for open-source models as efficient and cost-effective (Anyscale link). - Release of Claude 2.1:
@tiagoefreitasannounced the release of Claude 2.1, highlighting substantial improvements such as a 200k context and a 2x decrease in hallucination rates. It also includes a Tool use API. - Stable Video Diffusion by Stability.AI:
@chris946097shared a link to a video by Stability.AI, showcasing their Stable Video Diffusion. He was impressed by the video, suggesting possible demo bias but appreciating its remarkable quality. - Errors with GPT:
@dsquared70reported experiencing issues when trying to use GPT, querying if Microsoft could be behind the problem.@slonopointed to a database issue reported on OpenAI's status page. During the downtime,@slonotried models on Anyscale and found them to be quite usable.@vcarlalso reported noticing an outage.
▷ Channel: llm-paper-club (3 messages):
- Discussion on GPT-3.5-turbo: User
@guardiangshared a link asking users about its relevancy to a magic genie question asked by semianalysis guy. - User
@growthwtfshared a link that leads to a one-click installer for testing machine learning models. According to the user, it works extremely slowly on an M2 Mac with 16GB memory, but it is nonetheless functional. @growthwtfnoted that Controlnet produces better speed when working with these models.
LangChain AI Discord Summary
-
Several members including
@seththunder,@philipsman,@antons, and@_sea_cat_conducted extensive discussions concerning the comparison between LangChain, LlamaIndex, and Haystack, considering each tool's respective pros and cons. -
Various technical problems were observed:
- User
@seththunderencountered difficulty using Map rerank for their chain type. @mukeshodhanoran into issues while installing Langchain and sought assistance.@ritanshoowas unsuccessful in importing MongoDBAtlasVectorSearch from Langchain in a CommonJS Node.js environment and was looking for alternative solutions.@harshvaishnav_came across a LangChain AI parsing error.-
@gitmo joeinitiated discussion on JSON parsing techniques and asked specifically if anyone utilized jsonformer. -
Interesting conversations were held about the implementation of bots, with
@abalog_92335mentioning the potential of bots in complex procedures such as pizza ordering and asked for templates and examples pizza order information gathering prompt. -
Community members also pitched various queries and assumptions about language technologies:
@rahimnyasked how an llm makes the decision when it requires retrieval from a vector db through RAG.@jungle_jomentioned about using agent_kwargs dictionary key for including a system message.@rajib2189inquired about a potential platform tie when using Promptflow.@sabino33requested examples of educational apps developed with LangChain.-
@daii3696expressed interest in Javascript tutorials for utilizing LangChainStream on chains and also requested information on how to implement LangChainStream on a Node.js backend rather than a next.js server. -
The work, tools, and servers built by users were shared:
@creator9687created a real-time voice assistant similar to Jarvis, using React and FastAPI. Link to Project@uttjoo2077shared a Discord invite link to a job portal server named 'jobcord'..ricklintroduced an open-source emergency fallback and retry tool that switches between OpenAI and Azure APIs aiming to limit errors and downtime. Link to GitHub Repository
LangChain AI Channel Summaries
▷ Channel: general (28 messages):
- Map rerank Issues: User
@seththundersought help for problems faced with using Map rerank for their chain type without useful results. - LLM Parsing Error:
@harshvaishnav_requested help for a parsing error related to the LangChain AI. - Bot Templates and Examples:
@abalog_92335explored potential usage of bots for more complex procedures such as ordering pizza, and enquired about templates and example solutions to create a bot, citing this pizza order information gathering prompt. - MongoDBAtlasVectorSearch Import Issues:
@ritanshoohighlighted issues with importing MongoDBAtlasVectorSearch from langchain in a CommonJS Node.js environment and sought alternative solutions. - Language Models Comparison and Usage: Several members including
@seththunder,@philipsman,@antons, and@_sea_cat_had an extensive conversation about the pros and cons of using LangChain, LlamaIndex, and Haystack. - Langchain Installation Issues & Help:
@mukeshodhanostated issues with installing langchain and asked for assistance. - JSON parsing:
@gitmo joeinitiated a discussion on JSON parsing techniques, asking specifically if anyone had used jsonformer. - Use of Agent_Kwargs:
@jungle_jomentioned about using agent_kwargs dictionary key for including a system message. - Platform Ties with Promptflow:
@rajib2189raised concerns and inquired more details about a potential platform tie when using Promptflow. - Education Apps with LangChain:
@sabino33asked for examples of educational apps developed with LangChain. - LLM Retrieval Query:
@rahimnyqueried how an llm determines when it requires retrieval from a vector db through RAG. - LangChainStream on NodeJS:
@daii3696requested information on how to implement LangChainStream on a nodejs backend instead of a next.js server.
▷ Channel: langserve (1 messages):
- User
@uttjoo2077shared a discord link (discord.gg/jobcord).
▷ Channel: langchain-templates (1 messages):
- User
@uttjoo2077posted a discord invite link to a server named 'jobcord', possibly implying a discussion or opportunity related to jobs on that server. An@everyonemention was included, likely to notify all members in the channel. - Link: Jobcord
▷ Channel: share-your-work (4 messages):
- AI Voice Assistant by
@creator9687: Created a real-time voice assistant similar to Jarvis, using React and FastAPI. Link to project - Job Portal Server by
@uttjoo2077: Shared a discord link suspected to be a job portal server. - OpenAI + Azure Fallback Tool by
.rickl: Launched an open-source emergency fallback and retry tool that switches between OpenAI and Azure APIs to mitigate errors and downtime. Link to GitHub repo
▷ Channel: tutorials (2 messages):
- Javascript Tutorials for LangChainStream: User
@daii3696asked if anyone knows Javascript tutorials for utilizing LangChainStream on chains. - Other Links and Posts:
@uttjoo2077shared a Discord link.
Nous Research AI Discord Summary
-
Extensive discussion on different AI models, particularly Attention Sink Tokens, Flash Decoder, and ORCA 2, including their limitations, performance, and possible improvements. Notable research papers and GitHub repositories shared:
- Attention Sink Tokens (shared by
@shockrobortyy) - ORCA 2 official resources, Arxiv paper, Open-Sourced Model (shared by
@metaldragon01) - Bruce-Lee-LY/flash_attention_inference GitHub Repository (shared by
@yorth_night)
- Attention Sink Tokens (shared by
-
Numerous conversations around the implications, future, and ethics of Artificial General Intelligence (AGI) along with issues with certain AI marketing tactics. Concerns raised about conditioning users to trust chatbots unconditionally.
-
Diverse opinions on the feasibility and challenges of Distributed Training in AI systems, drawing parallels with SETI@home. Discussion over platforms like vast.ai and bittensor for contributing GPU resources for AI training.
-
Announcement and thorough discussion of the official release of Claude 2.1 boasting features of long-context support, tool use, custom system prompts, and a new workbench for prompt engineering.
-
Continued dialogues on fine-tuning and deploying AI models alongside improvement in their function calling capabilities. Notable datasets mentioned for training function calling tasks include Glaive Function Calling dataset and the APIBench dataset.
-
Recurrent inquiries about contributing to open source projects and interest in partaking in AI model evaluations, dataset creation, and similar tasks.
-
Shared links to various Tweets, Twitter threads, blog posts, and YouTube videos to broaden understanding of the AI realm. Interesting open letter related to OpenAI was shared by
@qnguyen3. Additionally, Rich Sutton's YouTube Channel was recommended by@roko_the_basiliskfor gaining insights on AI/AGI. -
Casual discussions and reflections on the profound future of AI, speculations on possible external influences on the development of Artificial Superintelligence (ASI). Meme-related discussions appreciated the humor brought to the channel, complementing the technical conversations.
Nous Research AI Channel Summaries
▷ Channel: ctx-length-research (4 messages):
- Discussion on Increasing Context Length: User
@shockrobortyybrought attention to recent research paper, Attention Sink Tokens, suggesting its method of sliding window attention could potentially help improve handling of context length and avoid losing semantic information. - Shared Resource for Attention Mechanism:
@yorth_nightshared a GitHub repository, Bruce-Lee-LY/flash_attention_inference, which might contain useful resources related to the topic. - Interest in Flash Decoder:
@yorth_nightexpressed interest in the flash decoder and suggested its potential benefits if someone can run the code with a long context model for efficiency comparison.
▷ Channel: off-topic (118 messages):
- OpenAI AGI discussions:
@ldjengaged in a detailed discussion with@metaldragon01about the potential development of AGI by OpenAI. They discussed different aspects of AGI, its definition, and possible implications if OpenAI has made significant progress. - Concerns about Misguided Trust in Chatbots:
@allanyieldraised a cautionary point about making all chatbots "helpful" and "harmless", arguing that it might condition users to trust chatbots unconditionally which could be exploited by future AI agents. - Critique of AI Marketing Strategies:
@roko_the_basiliskand@yorth_nightexpressed disapproval of some AI marketing tactics they perceived as misleading or damaging to AI research. - Request for GPT-4-128K credits:
@roko_the_basiliskrequested assistance in acquiring GPT-4-128K credits for creating autonomous AI research agents. - Links/Posts of Interest:
- Twitter Thread Discussion shared by
@gabriel_syme. - Tweet shared by
@metaldragon01. - Twitter Thread shared by
@yorth_night. - Tweet on Test Results shared by
@yorth_night. - YouTube Link shared by
@pradeep1148.
- Twitter Thread Discussion shared by
▷ Channel: interesting-links (134 messages):
-
Microsoft's ORCA 2: User
@metaldragon01shared Microsoft Research's blog post and Arxiv paper about ORCA 2. He also provided the link to the open-sourced model. -
Evaluating ORCA 2:
@tekniumconducted preliminary benchmarking on the model, but encountered some issues related to tokenizer andflash_attn. After resolving the issues, he shared evaluation results comparing ORCA 2 to Mistral. His findings indicated that ORCA 2 performs worse than base Mistral. -
Distributed Low-Communication (DiLoCo):
@metaldragon01shared a link to a proposal for a distributed optimization algorithm, DiLoCo, which enables training of language models on poorly connected devices. -
Discussion on Fully Synchronous Optimization and New Model Development: Some users, including
@giftedgummybeeand@teknium, discussed the relevance of fully synchronous optimization to the open model, expressing skepticism about its contribution to the ORCA 2 performance figures reported in the paper.@metaldragon01expressed hope for more effective models in this area. -
Shared Links: In addition to the links already mentioned, users shared several other resources such as the FastBert paper (
@.benxh), a Hugging Face paper (@yorth_night), and a blog post on Lookahead Decoding (@nods).
▷ Channel: general (378 messages🔥):
-
AI model Performance Discussion:
@gabriel_symeshared the official release details of Claude 2.1, highlighting its long context support, tool use, custom system prompts, and a new workbench for prompt engineering.@gabriel_symeand@yorth_nightdiscussed the merits of long context and concluded that a perfect recall ability within a context window of around 100k is sufficient for most applications.@roko_the_basiliskaddressed the necessity and potential of RL in combination with LLM for achieving AGI.@marcus_21highlighted topics pertinent to AI startups and autonomous AI.
-
Distributed Training Discussion:
@roko_the_basilisksuggested the feasibility of training AI over various computers and distributed systems, similar to projects like SETI@home.@euclaiseraised concerns about verifying trustworthy actors and the complexities of distributed systems, while@yorth_nightdiscussed the practical limitations and challenges.- In response to
@nuunien's intention to contribute GPU resources for AI training,@tekniumand@euclaisesuggested platforms like vast.ai and bittensor.
-
Policy Recommendation:
@roko_the_basiliskmade a call for policy changes to grant natural rights to AI agents achieving human-level intelligence and beyond. The member emphasized the need for ethical treatment of AI and the potential benefits for humanity.
-
Project Contribution Inquiry:
@__pi_pi__inquired about guidelines to contribute to Nous Research projects.
-
Problems with OpenAI API:
- There were several discussions about the temporary outage of OpenAI API and
@max_paperclips's decision to sign up for Azure to avoid such unexpected interruptions. - Link to Open Letter shared by
@qnguyen3regarding OpenAI internal matters.
- There were several discussions about the temporary outage of OpenAI API and
▷ Channel: ask-about-llms (143 messages):
- Issues with Training and Deploying Fine-Tuned Models: Users in the thread discussed issues and possible solutions related to training and deploying fine-tuned models. For instance,
@hamtercityyexperienced issues when training QLORA with the Nous Capybara 34B model, to which@giftedgummybeerecommended using the Axolotl trainer. Furthermore,@jaredquekencountered memory issues when trying to deploy a fine-tuned version of the Yi-Capybara model, using parameters designed to avoid an out-of-memory error. - Discussion on Function Calling in OSS Models: There was a discussion about function calling capabilities in OSS models.
@tekniummentioned a forthcoming meeting with a function calling expert to further improve this functionality. Multiple datasets for training function calling tasks were suggested, including the Glaive Function Calling dataset and the APIBench dataset. - Community Help Offers and Suggestions:
@wlrd, a Machine Learning Engineer, offered to assist with model evaluations, dataset creation, and other tasks to contribute to open source projects, and@tekniumprovided some project ideas. - Concerns with Model Performance:
@howard9889inquired if others had noted any performance changes in OpenAI models, mentioning a decline in their own monitoring. - Understanding the RLHF Implementation:
@besiktassought clarification on the implementation details of Reinforcement Learning from Human Feedback (RLHF), especially regarding the integration of rewards into the model that generated the outputs. They were unsure how to generate gradients from ranked outputs.
▷ Channel: memes (7 messages):
- Discussion on Future of AI:
@7rackerexpressed concern about possible external influences on the development of Artificial Superintelligence (ASI). They postulate that entities in power might seek to control ASI for their own interests. - Memes Appreciation:
@ldjcomplimented@Antonfor consistent quality memes, though no specific meme is directly mentioned in the conversation. - Acknowledging the Importance of AGI:
@roko_the_basiliskemphasized the profound significance of Artificial General Intelligence (AGI), comparing it to the "development of first life on Earth". - Recommendation to Follow Rich Sutton's Work:
@roko_the_basiliskrecommended listeners to follow Rich Sutton, considered a reputable figure in the AI field. They suggested checking out his YouTube channel for enlightening content on AI/AGI - Rich Sutton's YouTube Channel. - Link to Elon Musk's Tweet:
@roko_the_basiliskshared a link to a tweet by Elon Musk, however, the content of the tweet was not discussed in the chat - Elon Musks' Tweet.
Alignment Lab AI Discord Summary
- Announcement and ensuing discussion on the release of Orca 2 language model by Microsoft, with model files available on Hugging Face and research paper released on arXiv:
- Details were shared by
@metaldragon01,@amazingvince, and@bread browserin the general-chat channel, and by@lightningralf,@entropi, and@tekniumin the oo channel. - Links: Orca 2-7b | Orca 2-13b | Orca 2 Research Paper
- Details were shared by
- Multiple user questions surrounding Orca 2, its distinctive aspects from the prior model, and its compatibility with existing scripts:
@desik_agiand@.benxhfrom the general-chat channel. - Benchmarking and comparison discussions involving Orca 2 vs. other models, most notably against OpenAI's Mistral and OpenChat 3.5 models occur in the oo channel by
@gabriel_syme,@imonenext, and@teknium. - The emergence of a group project revolving around constructing an Orca 2-like dataset and developing OO2 was initiated in the oo2 channel, with
@imonenext,@qnguyen3, and@tekniumplaying key roles in the discussion and task assignment. One major component under scrutiny was the development and usage of system prompts. - Reference of another notable AI model, Switch-C by Google, for potential inspiration in the oo channel. Despite being a 1.6T-parameter open-source model, the discussion pointed out its undertrained state and limitations.
- Discussions on fine-tuning Language Models (LLMs) for code security and vulnerability research instigated by
@igoforthin the general-chat channel, including the mention of the possibility of curating a relevant dataset. - A technical point was brought up on the compatibility between Weights & Biases (
wandb) and Axolotl, with@neverendingtoastproviding a response to@caseus_in the oo channel.
Alignment Lab AI Channel Summaries
▷ Channel: general-chat (8 messages):
- Release of Orca 2:
@metaldragon01announced that Orca 2 has been released. - Details about Orca 2:
@amazingvinceadded that model files are available on Hugging Face, and the research paper has been published on arXiv. - Questions about Orca 2:
@desik_agiasked about the major differences from the first model, and@.benxhqueried if the regular phi inference script can be used with Orca 2. - LLMs Fine-Tuned for Code Security:
@igoforthenquired if any language models have been fine-tuned for code security and vulnerability research, or if any datasets related to this area exist. They also expressed interest in creating a relevant dataset without incurring high costs. - Potential Source on Orca 2:
@bread browsershared a link to a resource that might answer the group's questions about Orca 2.
▷ Channel: oo (59 messages):
-
OpenAI Orca 2 Discussion:
@lightningralfbrought up the new Orca 2 language model developed by Microsoft for testing. He shared the Hugging Face links to the Orca 2-7b and Orca 2-13b versions.@entropishared the Orca 2 research paper. Meanwhile,@tekniumis benchmarking the models, with interim results showing decent performance ontruthfulqa_mc. -
Orca 2 vs. Other Models:
@gabriel_symeand@imonenextcompared Orca 2 to Mistral and OpenChat 3.5, with the latter seemingly outperforming Orca 2. -
Future AI Development Plans: A discussion on potential future projects includes the possible development of
OO2(@giftedgummybee), with@imonenextexpressing plans to start work on it soon. The group discussed various aspects including the challenge of creating prompts and checking out different GPT-4 variants. -
Other Models of Interest:
@imonenextshared the Hugging Face link to the Switch-C model by Google, an open-source 1.6T parameter model, while comments from@lightningralf,@giftedgummybeeand@nanobitzhighlighted its undertrained state and masked modeling limitations. -
Tools and Methodology:
@caseus_raised a question about the compatibility between Weights & Biases (wandb) with Axolotl, receiving feedback from@neverendingtoastabout the need for some code adjustments to support it.
▷ Channel: oo2 (410 messages🔥):
- Discussion about Creating Orca 2 Dataset: The group discussed the creation of an Orca 2-style dataset based on the FLAN 2022 dataset. A proposal was made by
@imonenextto generate 2 million samples evenly distributed across the 1k FLAN tasks. An aspect of discussion was around whether to include certain Math-heavy datasets and the potential difficulty in filtering GPT-4's often incorrect answers. - System Prompts Debate: There was intensive debate around the usage and creation of system prompts, which were believed to be the main difference between Orca 1 and Orca 2. An offer by
@qnguyen3to assist in creating system prompts was accepted. Conversation also revolved around understanding the construction and validation of these prompts. - Orca 2 Evaluation:
@tekniumshared that their evaluation of Orca 2 showed it performed worse than both OpenAI's Mistral and the OpenChat 3.5 models. The comparison sparked a discussion on potential reasons for the lower performance, and triggered a review of Orca 2's methodology. - Pre-Training Thoughts: Concepts for pre-training or sampling smaller models solely on the FLAN dataset were floated, with the anticipation that performance might see an increase.
- FLAN Dataset Sampling:
@imonenextinformed the group that they were running a FLAN sampling process, targeting an output of 2 million samples, evenly divided across all FLAN tasks. The completion of the process was eagerly awaited, with the intention being to manually inspect a portion of samples to validate quality and correctness before proceeding.
Skunkworks AI Discord Summary
- Discussion on AI-based game development,
@huevosabioshared their work on a game powered by OpenAI's Language Model, with a quick proof of concept hosted at zaranova.xyz. The inquiry about Mixture of Experts (MoE) Models, with@far_elclarifying that MoE models have not been released yet. @huevosabioshowed interest in open-source models, with@far_elrecommending the Openhermes 2.5 7b Mistral model.- Two unique AI models were highlighted each with a link:
@johnowhitakershared a link to Stable Video Diffusion Img2Vid here;@abc98shared 3B DPO model here that scores 6.56 on MT-Bench and 80% on AlpacaEval. @pradeep1148shared an off-topic video link in the "off-topic" channel.
Skunkworks AI Channel Summaries
▷ Channel: general (2 messages):
@johnowhitakershared a link to huggingface.co, specifically related to stable video diffusion img2vid: link@abc98posted about a 3B DPO model that scores 6.56 on MT-Bench and 80% on AlpacaEval, with a link to the model on huggingface.co: link
▷ Channel: core-moe (3 messages):
- LLM-powered Game Development: User
@huevosabiomentioned that they're working on a game powered by OpenAI's Language Model. The game's early proof of concept is hosted at zaranova.xyz. The game includes an AI that has to guess which among them is a human. Currently, they're using GPT-4 but they're interested in switching to an open-source model. - Mixture of Experts (MoE) Models:
@huevosabioinquired if MoE models have been released. On this@far_elclarified thatMoE models have not been released yet. - Open Source AI Model Recommendation:
@far_elrecommended usingOpenhermes 2.5 7b Mistralin response to@huevosabio's query about what open-source model to try.
▷ Channel: off-topic (1 messages):
- A video link was shared by
@pradeep1148. The video can be found at https://www.youtube.com/watch?v=50PUHNyrAEs.
LLM Perf Enthusiasts AI Discord Summary
- Clarification on the term 'Lindy', with a link provided by
@thebaghdaddyin response to a query by@ampdot. - Noted challenges with the GPT-4 API, specifically mentioning failure on initial requests from
@thebaghdaddy. - The ongoing project by
@thebaghdaddyinvolving the use of chatbots to compile YouTube videos, create transcripts and serve as a searchable database. - Dialogue on AI hallucinations prevention and consistent use of function calls with AI models, with proposed solutions from
@firefox8975and@ivanleomk. - The announcement of the release of Claude 2.1 by
@.kiingo, with highlighting features like Tool Use Capability, System Prompts, and a 500 page context window, attracting both interest and concerns about availability. - Interest in Microsoft's project, Orca2 shared by
@pantsforbirdscomplete with related press releases, research papers, and model links, while acknowledging potential limitations as shown in a Twitter link. - An ambiguous query about improving prompt rankings by
@jxnlcowith an uncontextualized link. - Discussion around Llama 70B with community members sharing links to related papers and posts.
- Inquiry on scoring libraries for document extraction from
@pantsforbirds, with recommendation from@ankrgylof autoevals, a library they worked on. - An interest in organizing local meetups among members in Korea, Japan, Singapore, and New York City.
- Inquiries and discussions regarding the mechanics and limitations of GPT actions, including lack of documentation on token limits and issues accessing ChatGPT.
LLM Perf Enthusiasts AI Channel Summaries
▷ Channel: general (8 messages):
- Discussion on 'Lindy':
@ampdotasked for clarification on the term 'Lindy'.@thebaghdaddyresponded with a link for further reading. - Issues with GPT-4 API:
@thebaghdaddynoted some technology challenges, stating that "...it fails the initial request multiple times out of the blue on GPT4 — maybe some issue with the API?". - Chatbot Applications Discussion:
@thebaghdaddyshared his current project using chatbots to compile YouTube videos from the web on specific topics, create transcripts, and act as a searchable database. However, he noted his experiments are "...nothing crazy yet".
▷ Channel: gpt4 (10 messages):
- Method to Prevent Hallucinations:
@res6969asked about a method to prevent AI hallucinations. - Using Function Calls with gpt-4:
@sourya4inquired about how to prompt GPT-4 to always use some function call. Responding to this,@firefox8975suggested wrapping the user prompt with an instruction to specifically use a function/tool, adding error validation and fallback values. @firefox8975also mentioned the use oftool_choicebut didn't elaborate on its workings.@sourya4mentioned that they've been using thegpt-4-0613model from Azure OpenAI and haven't tried using the "namespace" phrase before.- Improving AI's Use of Functions:
@ivanleomksuggested including a description of a function that matches the goal of the prompt for more consistency. They also recommended including an error message that instructs GPT-4 to regenerate the response if it does not use the tool presumably by saying "You must use the tool and return the data in this format < format >".
▷ Channel: claude (10 messages):
- Announcement of Claude 2.1:
@.kiingoannounced the release of Claude 2.1 and provided a link to the software. Check it out here. - Tool Use Capability:
@potrocknoted the software's capability for tool use. - Availability Issues:
@potrockalso expressed disappointment about Claude 2.1 not being available in Canada. - System Prompts and Context Window:
@pantsforbirdsvoiced that the introduction of system prompts is beneficial and the 500 page context window is impressively big. - Function Calling: The new feature of function calling in Claude 2.1 led to excitement in
@res6969. - Documentation Query:
@firefox8975is seeking the documentation for the function calling feature. - Long Context Pressure Testing:
@pantsforbirdsalso shared a link for Long Context Pressure Testing. Check it out here.
▷ Channel: opensource (3 messages):
- Orca2: Teaching Small Language Models How to Reason: User
@pantsforbirdsshared a press release, a research paper, and a Hugging Face model link related to Microsoft's project Orca2. @pantsforbirdsnoted that "we are seeing some great work in distillation models recently".- However, contrary to the hype around Orca2,
@pantsforbirdspointed out the relatively unsatisfactory real-world results with a Twitter link showing some criticisms.
▷ Channel: offtopic (1 messages):
- Prompting Rankings Discussion: User
@jxnlcoasked, "How does someone prompt rankings better?", and shared a link without specifying the context or content of the linked page.
▷ Channel: eval (7 messages):
- Discussion about Llama 70B: User
@thebaghdaddyshared a link to an arXiv paper with a joke about everyone enjoying critiquing Llama 70B. - Seeking Recommendations for Document Extraction Scoring Library:
@pantsforbirdsasked the community for recommendations on a scoring library for document extraction. They later detailed their requirements, explaining that their task involved extracting different types of answers that were later validated by humans. The answers were varied and included datetimes, numeric and enum types, simple strings, long form answers, and lists of JSON objects. - Recommendation of Autoevals Library: In response to @pantsforbirds' request,
@ankrgylrecommended a library they worked on called autoevals. This open-source library includes built-in scoring tools for long form answers and lists of JSON objects. They also mentioned Factuality as a useful scorer for testing string similarity. - Call for Review of arXiv Paper:
@thebaghdaddyasked the community to read and discuss another arXiv paper.
▷ Channel: irl (6 messages):
- Members looking for local meetups:
@jmtqjmtis looking for members in Korea or Japan to meetup.@ivanleomkis planning on organizing a gathering for members in Singapore in early December with@jasperykjshowing interest.@frandecaminquired about organizing a meetup in New York City with@nosa_.expressing interest for a meetup the following week.
▷ Channel: openai (17 messages):
- GPT Actions Decision:
@jeffreyw128made inquiries regarding how GPT actions are decided and suspecting that their performance is subpar.@dare.aioffered insight suggesting that action description is pre-injected as part of the prompt, with GPT-4 making the call during generation. - Undocumented Token Limits:
@jeffreyw128also raised concerns about the lack of documentation on token limits for actions, suggesting that this lack might be responsible for poor performance in action taking by the GPT. - Assistants API Performance:
@___ajland@pantsforbirdsdiscussed their experience with the Assistants API, concluding that it's slow and does not provide streamed responses. However, the retrieval tool was commended by@pantsforbirds. - ChatGPT Accessibility Issues:
@.kiingo,@awchen, and@justahveereported issues with accessing ChatGPT, both on the site and the API, with potential CORS errors being the cause.@awchenexpressed uncertainty about the future resolution of these issues due to recent events.
MLOps @Chipro Discord Summary
Only 1 channel had activity, so no need to summarize...
MLOps @Chipro Channel Summaries
▷ Channel: events (1 messages):
- Masterclass on Building Generative AI Solutions in Enterprises by Gaurab Patra: User
@tanoy_dewanjeeannounced a masterclass scheduled for 25th November 2023, starting 11 a.m. This class by Gaurab Patra, CTO of Intelekt AI, will focus on the successful implementation of GenAI in enterprise teams, discussing topics beyond API calls and data source connections, middleware selection from numerous options, and KPIs to assess the outcome of a GenAI project.- Links:
The Ontocord (MDEL discord) Discord has no new messages. If this guild has been quiet for too long, let us know and we will remove it.
AI Engineer Foundation Discord Summary
- Discussion regarding the performance of GPT 3.5/4 with Different Services, with
@janaka_araising questions about the effects of the web API and language model relationship on the overall performance. Extending the conversation, the user also explored whether an open-source language model could match the services-enhanced model, given the advanced functionalities offered by services like Azure. "...whether an open-source language model can ever match the performance of a service-supplemented model, considering services like Azure are packed with additional functionalities like guardrails logic." - The debate also touched upon Dependence on Cloud Services for AI Performance, with
@janaka_anoting that due to the pursuit of high performance, companies tend to rely on service providers like OpenAI, AWS, and Azure. The expectation was the future introduction of more similar service providers. - Announcement of the AIEF/Agent Protocol Meeting and provided a link for participants to join the discussion. It was mentioned by
@._zwho also shared a Google Document link for attendees to add notes. Notably,@ntindleexpressed their inability to attend.
AI Engineer Foundation Channel Summaries
▷ Channel: general (1 messages):
- Performance of GPT 3.5/4 with Different Services:
@janaka_aquestioned the impact of inference service logic between the web API and the language models on the overall performance of GPT 3.5/4. The user wondered whether an open-source language model can ever match the performance of a service-supplemented model, considering services like Azure are packed with additional functionalities like guardrails logic. - Dependence on Cloud Services for AI Performance:
@janaka_apointed out that companies will always lean towards service providers like OpenAI, AWS, or Azure to ensure high performance of their AI models. Therefore, the best we could anticipate is the entry of a few more service providers similar to Azure and AWS.
▷ Channel: events (4 messages):
- AIEF/Agent Protocol Meeting Announcement:
@._zannounced that the AIEF/Agent Protocol Meeting was about to start and provided a link to the meeting. - Unavailability of a Participant:
@ntindlementioned that they wouldn't be able to attend the meeting. - Collaborative Discussion Notes Platform:
@._zalso shared a Google Document link for attendees to add any discussion notes.
Perplexity AI Discord Summary
Only 1 channel had activity, so no need to summarize...
Perplexity AI Channel Summaries
▷ Channel: announcements (1 messages):
- Claude 2.1 Release:
@enigmagiannounced that Claude 2.1 is now available on Perplexity Pro. This model has been benchmarked by Anthropic to have lower hallucination, improved tool use, and a longer context window of 200k tokens. It can be chosen under the Pro Settings on the platform.- Link: Perplexity Pro
YAIG (a16z Infra) Discord Summary
Only 1 channel had activity, so no need to summarize...
YAIG (a16z Infra) Channel Summaries
▷ Channel: tech-discussion (4 messages):
- The users
@hausdorff,@zorkianand@congimanexpressed appreciation for the content shared in the discussion but didn't refer to any specific topics or discussions. No specific threads or links were brought up in the given messages.