VibeBuilders.ai Logo
VibeBuilders.ai

All Resources

Month of August in AI
reddit
LLM Vibe Score0
Human Vibe Score1
Difficult-Race-1188This week

Month of August in AI

🔍 Inside this Issue: 🤖 Latest Breakthroughs: This month it’s all about Agents, LangChain RAG, and LLMs evaluation challenges.* 🌐 AI Monthly News: Discover how these stories are revolutionizing industries and impacting everyday life: EU AI Act, California’s Controversial SB1047 AI regulation act, Drama at OpenAI, and possible funding at OpenAI by Nvidia and Apple.* 📚 Editor’s Special: This covers the interesting talks, lectures, and articles we came across recently. Follow me on Twitter and LinkedIn at RealAIGuys and AIGuysEditor to get insight on new AI developments. Please don't forget to subscribe to our Newsletter: https://medium.com/aiguys/newsletter Latest Breakthroughs Are Agents just simple rules? Are Agents just enhanced reasoning? The answer is yes and no. Yes, in the sense that agents have simple rules and can sometimes enhance reasoning capabilities compared to a single prompt. But No in the sense that agents can have a much more diverse functionality like using specific tools, summarizing, or even following a particular style. In this blog, we look into how to set up these agents in a hierarchal manner just like running a small team of Authors, researchers, and supervisors. How To Build Hierarchical Multi-Agent Systems? TextGrad. It is a powerful framework performing automatic “differentiation” via text. It backpropagates textual feedback provided by LLMs to improve individual components of a compound AI system. In this framework, LLMs provide rich, general, natural language suggestions to optimize variables in computation graphs, ranging from code snippets to molecular structures. TextGrad showed effectiveness and generality across various applications, from question-answering and molecule optimization to radiotherapy treatment planning. TextGrad: Improving Prompting Using AutoGrad The addition of RAG to LLMs was an excellent idea. It helped the LLMs to become more specific and individualized. Adding new components to any system leads to more interactions and its own sets of problems. Adding RAG to LLMs leads to several problems such as how to retrieve the best content, what type of prompt to write, and many more. In this blog, we are going to combine the LangChain RAG with DSPy. We deep dive into how to evaluate the RAG pipeline quantitatively using RAGAs and how to create a system where instead of manually tweaking prompts, we let the system figure out the best prompt. How To Build LangChain RAG With DSPy? As the field of natural language processing (NLP) advances, the evaluation of large language models (LLMs) like GPT-4 becomes increasingly important and complex. Traditional metrics such as accuracy are often inadequate for assessing these models’ performance because they fail to capture the nuances of human language. In this article, we will explore why evaluating LLMs is challenging and discuss effective methods like BLEU and ROUGE for a more comprehensive evaluation. The Challenges of Evaluating Large Language Models AI Monthly News AI Act enters into force On 1 August 2024, the European Artificial Intelligence Act (AI Act) enters into force. The Act aims to foster responsible artificial intelligence development and deployment in the EU. The AI Act introduces a uniform framework across all EU countries, based on a forward-looking definition of AI and a risk-based approach: Minimal risk: most AI systems such as spam filters and AI-enabled video games face no obligation under the AI Act, but companies can voluntarily adopt additional codes of conduct. Specific transparency risk: systems like chatbots must clearly inform users that they are interacting with a machine, while certain AI-generated content must be labelled as such. High risk: high-risk AI systems such as AI-based medical software or AI systems used for recruitment must comply with strict requirements, including risk-mitigation systems, high-quality of data sets, clear user information, human oversight, etc. Unacceptable risk: for example, AI systems that allow “social scoring” by governments or companies are considered a clear threat to people’s fundamental rights and are therefore banned. EU announcement: Click here https://preview.redd.it/nwyzfzgm4cmd1.png?width=828&format=png&auto=webp&s=c873db37ca0dadd5b510bea70ac9f633b96aaea4 California AI bill SB-1047 sparks fierce debate, Senator likens it to ‘Jets vs. Sharks’ feud Key Aspects of SB-1047: Regulation Scope: Targets “frontier” AI models, defined by their immense computational training requirements (over 10²⁶ operations) or significant financial investment (>$100 million). Compliance Requirements: Developers must implement safety protocols, including the ability to immediately shut down, cybersecurity measures, and risk assessments, before model deployment. Whistleblower Protections: Encourages reporting of non-compliance or risks by offering protection against retaliation. Safety Incident Reporting: Mandates reporting AI safety incidents within 72 hours to a newly established Frontier Model Division. Certification: Developers need to certify compliance, potentially under penalty of perjury in earlier drafts, though amendments might have altered this. Pros: Safety First: Prioritizes the prevention of catastrophic harms by enforcing rigorous safety standards, potentially safeguarding against AI misuse or malfunction. Incentivizes Responsible Development: By setting high standards for AI model training, the company encourages developers to think critically about the implications of their creations. Public Trust: Enhances public confidence in AI by ensuring transparency and accountability in the development process. Cons: Innovation Stagnation: Critics argue it might stifle innovation, especially in open-source AI, due to the high costs and regulatory burdens of compliance. Ambiguity: Some definitions and requirements might be too specific or broad, leading to legal challenges or unintended consequences. Global Competitiveness: There’s concern that such regulations could push AI development outside California or the U.S., benefiting other nations without similar restrictions. Implementation Challenges: The practicalities of enforcing such regulations, especially the “positive safety determination,” could be complex and contentious. News Article: Click here Open Letter: Click here https://preview.redd.it/ib96d7nk4cmd1.png?width=828&format=png&auto=webp&s=0ed5913b5dae72e203c8592393e469d9130ed689 MORE OpenAI drama OpenAI co-founder John Schulman has left the company to join rival AI startup Anthropic, while OpenAI president and co-founder Greg Brockman is taking an extended leave until the end of the year. Schulman, who played a key role in creating the AI-powered chatbot platform ChatGPT and led OpenAI’s alignment science efforts, stated his move was driven by a desire to focus more on AI alignment and hands-on technical work. Peter Deng, a product manager who joined OpenAI last year, has also left the company. With these departures, only three of OpenAI’s original 11 founders remain: CEO Sam Altman, Brockman, and Wojciech Zaremba, lead of language and code generation. News Article: Click here https://preview.redd.it/0vdjc18j4cmd1.png?width=828&format=png&auto=webp&s=e9de604c26aed3e47b50df3bdf114ef61f967080 Apple and Nvidia may invest in OpenAI Apple, which is planning to integrate ChatGPT into iOS, is in talks to invest. Soon after, Bloomberg also reported that Apple is in talks but added that Nvidia “has discussed” joining the funding round as well. The round is reportedly being led by Thrive Capital and would value OpenAI at more than $100 billion. News Article: Click here https://preview.redd.it/ude6jguh4cmd1.png?width=828&format=png&auto=webp&s=3603cbca0dbb1be3e6d0efcf06c3a698428bbdd6 Editor’s Special The AI Bubble: Will It Burst, and What Comes After?: Click here Eric Schmidt Full Controversial Interview on AI Revolution (Former Google CEO): Click here AI isn’t gonna keep improving Click here General Intelligence: Define it, measure it, build it: Click here

Need Advice on Implementing Reranking Models for an AI-Based Document-Specific Copilot feature
reddit
LLM Vibe Score0
Human Vibe Score1
Swimming_Teach_7579This week

Need Advice on Implementing Reranking Models for an AI-Based Document-Specific Copilot feature

Hey everyone! I'm currently working on an AI-based grant writing system that includes two main features: Main AI: Uses LLMs to generate grant-specific suggestions based on user-uploaded documents. Copilot Feature: Allows document-specific Q&A by utilizing a query format like /{filename} {query} to fetch information from the specified document. Currently, we use FAISS for vector storage and retrieval, with metadata managed through .pkl files. This setup works for similarity-based retrieval of relevant content. However, I’m considering introducing a reranking model to further enhance retrieval accuracy, especially for our Copilot feature. Challenges with Current Setup: Document-Specific Retrieval: We're storing document-specific embeddings and metadata in .pkl files, and retrieval works by first querying FAISS. Objective: Improve the precision of the results retrieved by Copilot when the user requests data from a specific document (e.g., /example.pdf summarize content). Questions for the Community: Is using a reranking model (e.g., BERT-based reranker, MiniLM) a good idea to add another layer of precision for document retrieval, especially when handling specific document requests? If I implement a reranking model, do I still need the structured .pkl files, or can I rely solely on the embeddings and reranking for retrieval? How can I effectively integrate a reranking model into my current FAISS + Langchain setup? I’d love to hear your thoughts, and if you have experience in using reranking models with FAISS or similar, any advice would be highly appreciated. Thank you!

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.
reddit
LLM Vibe Score0
Human Vibe Score0.714
level6-killjoyThis week

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.

This is a recap covering the major news from last week. 🔥Top 3 news - OpenAI’s updates, Meta’s upcoming free LLM and EU Regulation 🗞️Interesting reads include PSA about protecting your keys, The GPT ouroboros, Reddit - OpenAI’s moat, and more.. 🧑‍🎓Learning includes a Step-by-step guide from a non-technical founder who launched his MVP, Chatbot for your Gdrive and more 🔥Top 3 AI news in the past week OpenAI: New Pricing, Models, & Functions OpenAI has been on a roll. Last week we saw the release of OpenAI best practice on using GPT. This week we saw some amazing updates. Three major buckets were: First, the price decreases for both embeddings and GPT-3.5 tokens. Second, new models for gpt-4 and gpt-3.5. A new longer context model for gpt-3.5. Third, a new function calling capability. Why is it important? Previously, the output from OpenAI was all text. So, calling an external API from GPT was quite difficult. You had to parse the text data and things were often incorrect. Langchain created the Agents and Tools feature to tackle this problem. It was still unreliable and prone to issues. Now you get native support to generate a fixed format output. You can use the output to generate functional calls and also pass functions which need to be called. For example, if your app has multiple API endpoints then you can use GPT to generate the API calls with parameters. You can also pass the endpoints as function calls to ensure the correct function is executed. This functionality can further be used to generate structured data (JSON) out of GPT. So, you can generate data from GPT and load it into your backend. What’s next? This functionality allows turning natural language responses into structured data. This can be used to create “intelligent” backends using LLMs. We might see implementations in no-code tools to allow more robust and natural-language tools for non-technical folks. The structured data process goes both ways. You can also feed structured data into GPT for better responses. This feature also has its share of issues. Function calling suffers from the same prompt injection issues. Malicious actors can pass malicious code in function or the responses. For example, creation of queries using functions might contain malicious code to delete data. Without proper user validation this code will be executed automatically and delete data. So, using LLM as the back-end layer needs proper security implementation. Meta's LLM: Commercial Use Ahead Llama has been a boon for the open source community. Many of the open source models rely on Llama. The issue is that Llama is research-only and cannot be used commercially. So, no one can use it to build any product. Meta is now working on the next version of the model. This model will be available for commercial use. This is in stark contrast to both OpenAI and Google. Both safe-guarde their models and make it available through API. Why is it important? Certain industries cannot use LLM APIs because of strict restrictions on data privacy. These companies would want to run their own instance of a foundational model. A commercially available foundational model is also going to help people who want to keep their “API call” costs next to 0. A commercially available free-for-all model will also help push the open source community further. Just like Llama. What’s next? Sam Altman has said OpenAI didn’t release GPT-3 as open-source because they didn’t think people would be able to run it. Now OpenAI is working on an open-source model. This is going to be weaker than GPT-4. Let the battle of LLMs begin. EU's Proposed Legislation and Its Impact on AI Usage The EU parliament voted to move ahead with the E.U. AI Act. This act aims to ensure consumer protection against the dangers of AI. Why is it important? OpenAI and Sam Altman want regulations for models. They have proposed a IAEA-type of agency to stop the proliferation of LLM models. As per OpenAI, all models should be regulated and monitored. The suggestion of a license based regulation has led to significant backlash. Many people have called it “regulatory capture” - with the aim of shutting down competing LLMs. Licensing based regulations might not really be effective. The EU is approaching regulation from a different angle. It doesn’t focus on how models are developed. Rather focuses on how AI will/can be used. They have broken down use cases into 4 categories - unacceptable (prohibited), high, medium and low risk. For example, Building a Pre-Crime software,on%20crimes%20not%20yet%20committed.) to predict crimes? Building a Social credit system? Unacceptable. Using tools to influence elections or recommendation algorithms? High (Highly regulated). Using generative AI tools to create text or images on news sites? Medium (Add label that the content is AI generated) AI providers also need to disclose their training source. To me this sounds like good legislation. What do you guys think? But, OpenAI has warned that EU regulations might force them to pull out completely. What’s next? The disclosure requirements might help various publishing companies. AI and media companies are in talks to pay for training data. Google has been leading the charge. Additionally, OpenAI and Deepmind will open their models for safety and research purposes to the UK government. 🗞️10 AI news highlights and interesting reads PSA: If you are using Repl to write code, you might want to check your OpenAI API keys. If you have left them embedded then people can pirate and steal the keys. LLMs rely on human annotation or human feedback to learn. And one way to generate human annotation is crowdsourcing. But what if the crowdsource human annotators use LLMs? Research shows 33-46% workers used LLMs. So, basically we go from Human -> AI -> Human -> AI. The AI ouroboros. Researchers also say generated data to train models might cause serious issue. All the talks about moats \- Reddit might be OpenAI’s \future\ moat. Given the amount of complaints about how Google search experience has deteriorated during the blackout, this might be true? Doctors are using ChatGPT but not to diagnose.Rather to be more empathetic. We discussed this just a month ago. And guess where the data for this study came from? Reddit AskDocs. Moat FTW?! Beatles to make a comeback…using Generative AI. SnapFusion - Text to Image diffusion on mobile phones. Large context lengths are important for better GPT experience. The secret sauce for 100k context length. There is a lot of bad AI research out there. Some border on snake oil. Most AI “research” should be double checked and challenged. A new research on huggingface said that GPT-4 can ace MIT curriculum. Now someone is replicating the results and say that GPT-4 can’t beat MIT. Are we seeing peak AI? Especially when people from Deepmind and Meta are involved? Mistral AI raised $113 million in seed round with no product. Some might say this funding is for the team and the team is really solid. The issue though is whether the valuation is justified when OpenAI and Google already have a head start. The AI Hype Wall of Shame. \- Collection of articles which mislead people about AI in various aspects. 🧑‍🎓3 Learning Resources Building and Launching a company using GPT-4 with prompts. (The author didn’t know how to code but created and launched the MVP in a month). Chatbot for your Gdrive - https://www.haihai.ai/gpt-gdrive/ Building ChatGPT plugin using Supabase - https://supabase.com/blog/building-chatgpt-plugins-template That’s it folks. Thank you for reading and have a great week ahead. If you are interested in a focused weekly recap delivered to your inbox on Mondays you can subscribe here. It is FREE!

Join the AI4Earth challenge with the European Space Agency to highlight our footprint on Earth using Earth Observation data and Machine Learning
reddit
LLM Vibe Score0
Human Vibe Score1
campachThis week

Join the AI4Earth challenge with the European Space Agency to highlight our footprint on Earth using Earth Observation data and Machine Learning

​ https://preview.redd.it/ww109cba14f71.png?width=2401&format=png&auto=webp&s=8bd3d43e8b63848af85c73478be61e43d9e10189 The primary goal is to get an insight into the human impact on Earth, to drive and guide conservation efforts of this planet we call home. Our approach will be twofold:  Firstly we will work on AI algorithms that can serve as an early detection system of human impact sites. Secondly we will use these detection systems to find satellite images that show the most impactful human-caused changes, which will be used in the creation of a video to launch an awareness campaign. You will be working with ESA to detect things like: Wildfires and Deforestation Marine Litter and Melting Glaciers Air quality detection & Novel animal migration patterns  and much more!  European Space Agency To reach these goals we’ve partnered up with ESA, who are able to use our algorithms to monitor new satellite data and guide conservation efforts. They will provide us with multi-spectral data of their Sentinel-2 satellite pair and with invaluable knowledge and research on the domain of Earth Observation data in participant only masterclasses.  Format The challenge will run throughout September and October, where you will collaborate with a diverse team of over 30 international data specialists and domain experts in subteams, all tackling this problem from different angles. Subtasks like the detection of deforestation, wildfires, marine litter or any other human caused impact. All contributors in the challenge are expected to spend 12 hours or more per week during the entirity of the two month challenge. To learn more subscribe to the info session on the 3rd of August 19:00 CEST HERE! Some important dates: 3rd of August – Info session 1st of September – Challenge Kick-off 29th of September – Midterm presentations 29th of October – Final presentations PARTNERS SUN - https://spacehubs.network The project is spearheaded by SUN whose goal is to increase the commercialization of space enabled solutions and growth of European start-ups and scale-ups in the space downstream and upstream sectors. ESA - https://esa.int ESA will be the main stakeholder and domain knowledge provider in the challenge. Their efforts to aid human’s space endeavours as well as protect the planet we live on will serve us for many years to come.  MLReef - https://mlreef.com MLReef provides an open source platform for collaborative Machine Learning. They provide the computational infrastructure to support the EO4Earth project as part of their AI4GOOD and Open Science initiatives. Brimatech  As a partner in the SUN project, the innovation management and market research expert Brimatech helps out in the overall organisation of the challenge.  Mothership The ‘Mothership’ is a dedicated open innovation program created by Space4Good and World Startup Factory. The Mothershi is leveraging recent advancements in artificial intelligence and satellite technologies in support of the UN Sustainable Development Goals. Space4Good  Space4Good is a geospatial innovation lab supporting impact makers on the ground with earth observation insights from above. Worldstartup  Worldstartup is a collective of international entrepreneurs, experts, mentors and investors, dedicated to help the best impact-driven startups and scaleups.

Showing 481-504 of 1945 resources