VibeBuilders.ai Logo
VibeBuilders.ai

All Resources

Solana_AIAgent_Trading
github
LLM Vibe Score0.464
Human Vibe Score0.05777682403433476
solagent99Mar 25, 2025

Solana_AIAgent_Trading

Solana AI Agent Trading Tool An open-source trading toolkit for connecting AI agents to Solana protocols. Now, any agent, using any model can autonomously perform 15+ Solana actions: Trade tokens Launch new tokens Lend assets Send compressed airdrops Execute blinks Launch tokens on AMMs And more... 💬 Contact Me If you have any question or something, feel free to reach out me anytime via telegram, discord or twitter. 🌹 You're always welcome 🌹 Telegram: @Leo Replit template created by Arpit Singh 🔧 Core Blockchain Features Token Operations Deploy SPL tokens by Metaplex Transfer assets Balance checks Stake SOL Zk compressed Airdrop by Light Protocol and Helius NFTs on 3.Land Create your own collection NFT creation and automatic listing on 3.land List your NFT for sale in any SPL token NFT Management via Metaplex Collection deployment NFT minting Metadata management Royalty configuration DeFi Integration Jupiter Exchange swaps Launch on Pump via PumpPortal Raydium pool creation (CPMM, CLMM, AMMv4) Orca Whirlpool integration Manifest market creation, and limit orders Meteora Dynamic AMM, DLMM Pool, and Alpha Vault Openbook market creation Register and Resolve SNS Jito Bundles Pyth Price feeds for fetching Asset Prices Register/resolve Alldomains Perpetuals Trading with Adrena Protocol Drift Vaults, Perps, Lending and Borrowing Solana Blinks Lending by Lulo (Best APR for USDC) Send Arcade Games JupSOL staking Solayer SOL (sSOL)staking Non-Financial Actions Gib Work for registering bounties 🤖 AI Integration Features LangChain Integration Ready-to-use LangChain tools for blockchain operations Autonomous agent support with React framework Memory management for persistent interactions Streaming responses for real-time feedback Vercel AI SDK Integration Vercel AI SDK for AI agent integration Framework agnostic support Quick and easy toolkit setup Autonomous Modes Interactive chat mode for guided operations Autonomous mode for independent agent actions Configurable action intervals Built-in error handling and recovery AI Tools DALL-E integration for NFT artwork generation Natural language processing for blockchain commands Price feed integration for market analysis Automated decision-making capabilities 📃 Documentation You can view the full documentation of the kit at docs.solanaagentkit.xyz 📦 Installation Quick Start Usage Examples Deploy a New Token Create NFT Collection on 3Land Create NFT on 3Land When creating an NFT using 3Land's tool, it automatically goes for sale on 3.land website Create NFT Collection Swap Tokens Lend Tokens Stake SOL Stake SOL on Solayer Send an SPL Token Airdrop via ZK Compression Fetch Price Data from Pyth Open PERP Trade Close PERP Trade Close Empty Token Accounts Create a Drift account Create a drift account with an initial token deposit. Create a Drift Vault Create a drift vault. Deposit into a Drift Vault Deposit tokens into a drift vault. Deposit into your Drift account Deposit tokens into your drift account. Derive a Drift Vault address Derive a drift vault address. Do you have a Drift account Check if agent has a drift account. Get Drift account information Get drift account information. Request withdrawal from Drift vault Request withdrawal from drift vault. Carry out a perpetual trade using a Drift vault Open a perpertual trade using a drift vault that is delegated to you. Carry out a perpetual trade using your Drift account Open a perpertual trade using your drift account. Update Drift vault parameters Update drift vault parameters. Withdraw from Drift account Withdraw tokens from your drift account. Borrow from Drift Borrow tokens from drift. Repay Drift loan Repay a loan from drift. Withdraw from Drift vault Withdraw tokens from a drift vault after the redemption period has elapsed. Update the address a Drift vault is delegated to Update the address a drift vault is delegated to. Get Voltr Vault Position Values Get the current position values and total value of assets in a Voltr vault. Deposit into Voltr Strategy Deposit assets into a specific strategy within a Voltr vault. Withdraw from Voltr Strategy Withdraw assets from a specific strategy within a Voltr vault. Get a Solana asset by its ID Get a price inference from Allora Get the price for a given token and timeframe from Allora's API List all topics from Allora Get an inference for an specific topic from Allora Examples LangGraph Multi-Agent System The repository includes an advanced example of building a multi-agent system using LangGraph and Solana Agent Kit. Located in examples/agent-kit-langgraph, this example demonstrates: Multi-agent architecture using LangGraph's StateGraph Specialized agents for different tasks: General purpose agent for basic queries Transfer/Swap agent for transaction operations Read agent for blockchain data queries Manager agent for routing and orchestration Fully typed TypeScript implementation Environment-based configuration Check out the LangGraph example for a complete implementation of an advanced Solana agent system. Dependencies The toolkit relies on several key Solana and Metaplex libraries: @solana/web3.js @solana/spl-token @metaplex-foundation/digital-asset-standard-api @metaplex-foundation/mpl-token-metadata @metaplex-foundation/mpl-core @metaplex-foundation/umi @lightprotocol/compressed-token @lightprotocol/stateless.js Contributing Contributions are welcome! Please feel free to submit a Pull Request. Refer to CONTRIBUTING.md for detailed guidelines on how to contribute to this project. Contributors Star History License Apache-2 License Funding If you wanna give back any tokens or donations to the OSS community -- The Public Solana Agent Kit Treasury Address: Solana Network : EKHTbXpsm6YDgJzMkFxNU1LNXeWcUW7Ezf8mjUNQQ4Pa Security This toolkit handles private keys and transactions. Always ensure you're using it in a secure environment and never share your private keys.

ai-flow
github
LLM Vibe Score0.461
Human Vibe Score0.01809909681901274
DahnM20Mar 25, 2025

ai-flow

Open-source tool to seamlessly connect multiple AI model APIs into repeatable workflows. 🔗 Website • 📚 Documentation 🎉🚀 Latest Release: v0.10.0 🚀🎉 New Nodes: Claude 3.7, OpenRouter, Generate Random Number Configuration can now be done entirely in the UI !AI-Flow Intro Overview AI-Flow is an open-source, user-friendly UI that lets you visually design, manage, and monitor AI-driven workflows by seamlessly connecting multiple AI model APIs (e.g., OpenAI, StabilityAI, Replicate, Claude, Deepseek). Features Visual Workflow Builder: Drag-and-drop interface for crafting AI workflows. Real-Time Monitoring: Watch your workflow execute and track results. Parallel Processing: Nodes run in parallel whenever possible. Model Management: Easily organize and manage diverse AI models. Import/Export: Share or back up your workflows effortlessly. Supported Models Replicate: LLaMa, Mistral, FaceSwap, InstantMesh, MusicGen, and more. OpenAI: GPT-4o, TTS, o1, o3. StabilityAI: Stable Diffusion 3.5, SDXL, Stable Video Diffusion, plus additional tools. Others: Claude, Deepseek. !Scenario Example Open Source vs. Cloud AI-Flow is fully open source and available under the MIT License, empowering you to build and run your AI workflows on your personal machine. For those seeking enhanced functionality and a polished experience, AI-Flow Pro on our cloud platform (app.ai-flow.net) offers advanced features, including: Subflows & Loops: Create complex, nested workflows and iterate tasks effortlessly. API-Triggered Flows: Initiate workflows via API calls for seamless automation. Integrated Services: Connect with external services such as Google Search, Airtable, Zapier, and Make. Simplified Interface: Transform workflows into streamlined tools with an intuitive UI. !Pro VS Open Source The cloud version builds upon the foundation of the open-source project, giving you more power and flexibility while still letting you use your own API keys. Installation Note: To unlock full functionality, AI-Flow requires S3-compatible storage (with proper CORS settings) to host resources. Without it, features like File Upload or nodes that rely on external providers (e.g., StabilityAI) may not work as expected. Also, set REPLICATEAPIKEY in your environment to use the Replicate node. Local Installation (Without Docker) Clone the Repository: UI Setup: Backend Setup: Windows Users: Run the Application: Start the backend: In a new terminal, start the UI: Open your browser and navigate to http://localhost:3000. Docker Installation Prepare Docker Compose: Navigate to the docker directory: Update the REPLICATEAPIKEY in the YAML file. Launch with Docker Compose: Access the Application: Open http://localhost:80 in your browser. To stop, run: Contributing We welcome contributions! If you encounter issues or have feature ideas, please open an issue or submit a pull request. License This project is released under the MIT License.

aima-java
github
LLM Vibe Score0.521
Human Vibe Score0.06620214044837505
aimacodeMar 25, 2025

aima-java

AIMA3e-Java (JDK 8+) Java implementation of algorithms from Russell and Norvig's Artificial Intelligence - A Modern Approach 3rd Edition. You can use this in conjunction with a course on AI, or for study on your own. We're looking for solid contributors to help. Getting Started Links Overview of Project Interested in Contributing Setting up your own workspace Comments on architecture and design Demo Applications that can be run from your browser (unfortunately not up to date) Javadoc for the aima-core project (outdated) Download the latest official (but outdated) version = 1.9.1 (Dec 18 2016) Latest Maven Information (for integration as a third party library) Index of Implemented Algorithms |Figure|Page|Name (in 3rd edition)|Code | -------- |:--------:| :-----| :----- | |2|34|Environment|Environment| |2.1|35|Agent|Agent| |2.3|36|Table-Driven-Vacuum-Agent|TableDrivenVacuumAgent| |2.7|47|Table-Driven-Agent|TableDrivenAgentProgram| |2.8|48|Reflex-Vacuum-Agent|ReflexVacuumAgent| |2.10|49|Simple-Reflex-Agent|SimpleReflexAgentProgram| |2.12|51|Model-Based-Reflex-Agent|ModelBasedReflexAgentProgram| |3|66|Problem|Problem| |3.1|67|Simple-Problem-Solving-Agent|SimpleProblemSolvingAgent| |3.2|68|Romania|SimplifiedRoadMapOfRomania| |3.7|77|Tree-Search|TreeSearch| |3.7|77|Graph-Search|GraphSearch| |3.10|79|Node|Node| |3.11|82|Breadth-First-Search|BreadthFirstSearch| |3.14|84|Uniform-Cost-Search|UniformCostSearch| |3|85|Depth-first Search|DepthFirstSearch| |3.17|88|Depth-Limited-Search|DepthLimitedSearch| |3.18|89|Iterative-Deepening-Search|IterativeDeepeningSearch| |3|90|Bidirectional search|BidirectionalSearch| |3|92|Best-First search|BestFirstSearch| |3|92|Greedy best-First search|GreedyBestFirstSearch| |3|93|A\* Search|AStarSearch| |3.26|99|Recursive-Best-First-Search |RecursiveBestFirstSearch| |4.2|122|Hill-Climbing|HillClimbingSearch| |4.5|126|Simulated-Annealing|SimulatedAnnealingSearch| |4.8|129|Genetic-Algorithm|GeneticAlgorithm| |4.11|136|And-Or-Graph-Search|AndOrSearch| |4|147|Online search problem|OnlineSearchProblem| |4.21|150|Online-DFS-Agent|OnlineDFSAgent| |4.24|152|LRTA\*-Agent|LRTAStarAgent| |5.3|166|Minimax-Decision|MinimaxSearch| |5.7|170|Alpha-Beta-Search|AlphaBetaSearch| |6|202|CSP|CSP| |6.1|204|Map CSP|MapCSP| |6.3|209|AC-3|AC3Strategy| |6.5|215|Backtracking-Search|AbstractBacktrackingSolver| |6.8|221|Min-Conflicts|MinConflictsSolver| |6.11|224|Tree-CSP-Solver|TreeCspSolver| |7|235|Knowledge Base|KnowledgeBase| |7.1|236|KB-Agent|KBAgent| |7.7|244|Propositional-Logic-Sentence|Sentence| |7.10|248|TT-Entails|TTEntails| |7|253|Convert-to-CNF|ConvertToCNF| |7.12|255|PL-Resolution|PLResolution| |7.15|258|PL-FC-Entails?|PLFCEntails| |7.17|261|DPLL-Satisfiable?|DPLLSatisfiable| |7.18|263|WalkSAT|WalkSAT| |7.20|270|Hybrid-Wumpus-Agent|HybridWumpusAgent| |7.22|272|SATPlan|SATPlan| |9|323|Subst|SubstVisitor| |9.1|328|Unify|Unifier| |9.3|332|FOL-FC-Ask|FOLFCAsk| |9.6|338|FOL-BC-Ask|FOLBCAsk| |9|345|CNF|CNFConverter| |9|347|Resolution|FOLTFMResolution| |9|354|Demodulation|Demodulation| |9|354|Paramodulation|Paramodulation| |9|345|Subsumption|SubsumptionElimination| |10.9|383|Graphplan|GraphPlan| |11.5|409|Hierarchical-Search|HierarchicalSearchAlgorithm| |11.8|414|Angelic-Search|---| |13.1|484|DT-Agent|DT-Agent| |13|484|Probability-Model|ProbabilityModel| |13|487|Probability-Distribution|ProbabilityDistribution| |13|490|Full-Joint-Distribution|FullJointDistributionModel| |14|510|Bayesian Network|BayesianNetwork| |14.9|525|Enumeration-Ask|EnumerationAsk| |14.11|528|Elimination-Ask|EliminationAsk| |14.13|531|Prior-Sample|PriorSample| |14.14|533|Rejection-Sampling|RejectionSampling| |14.15|534|Likelihood-Weighting|LikelihoodWeighting| |14.16|537|GIBBS-Ask|GibbsAsk| |15.4|576|Forward-Backward|ForwardBackward| |15|578|Hidden Markov Model|HiddenMarkovModel| |15.6|580|Fixed-Lag-Smoothing|FixedLagSmoothing| |15|590|Dynamic Bayesian Network|DynamicBayesianNetwork| |15.17|598|Particle-Filtering|ParticleFiltering| |16.9|632|Information-Gathering-Agent|InformationGatheringAgent| |17|647|Markov Decision Process|MarkovDecisionProcess| |17.4|653|Value-Iteration|ValueIteration| |17.7|657|Policy-Iteration|PolicyIteration| |17.9|663|POMDP-Value-Iteration|POMDPValueIteration| |18.5|702|Decision-Tree-Learning|DecisionTreeLearner| |18.8|710|Cross-Validation-Wrapper|CrossValidation| |18.11|717|Decision-List-Learning|DecisionListLearner| |18.24|734|Back-Prop-Learning|BackPropLearning| |18.34|751|AdaBoost|AdaBoostLearner| |19.2|771|Current-Best-Learning|CurrentBestLearning| |19.3|773|Version-Space-Learning|VersionSpaceLearning| |19.8|786|Minimal-Consistent-Det|MinimalConsistentDet| |19.12|793|FOIL|FOIL| |21.2|834|Passive-ADP-Agent|PassiveADPAgent| |21.4|837|Passive-TD-Agent|PassiveTDAgent| |21.8|844|Q-Learning-Agent|QLearningAgent| |22.1|871|HITS|HITS| |23.5|894|CYK-Parse|CYK| |25.9|982|Monte-Carlo-Localization|MonteCarloLocalization| Index of implemented notebooks |Chapter No|Name |Status (in 3rd edition)|Status (in 4th edition) | -------- |:--------:| :-----| :----- | |3| Solving Problems by Searching| In Progress| Not started| |6| Constraint Satisfaction Problems |In Progress|---| |12| Knowledge Representation|Done|---| |13| Quantifying Uncertainty |Done | --- | |14| Probabilistic Reasoning|In Progress| ---| Before starting to work on a new notebook: Open a new issue with the following heading: Notebook: Chapter Name - Version . Check that the issue is not assigned to anyone. Mention a topics list of what you will be implementing in the notebook for that particular chapter. You can iteratively refine the list once you start working. Start a discussion on what can go in that particular notebook. "---" indicates algorithms yet to be implemented. Index of data structures Here is a table of the data structures yet to be implemented. |Fig|Page|Name (in book)|Code| | -------- |:--------:| :-----| :----- | |9.8|341|Append|---| |10.1|369|AIR-CARGO-TRANSPORT-PROBLEM|---| |10.2|370|SPARE-TIRE-PROBLEM|---| |10.3|371|BLOCKS-WORLD |---| |10.7|380|HAVE-CAKE-AND-EAT-CAKE-TOO-PROBLEM|---| |11.1|402|JOB-SHOP-SCHEDULING-PROBLEM|---| |11.4|407|REFINEMENT-HIGH-LEVEL-ACTIONS|---| |23.6|895|SENTENCE-TREE|---| |29.1|1062|POWERS-OF-2|---|

AI-PhD-S24
github
LLM Vibe Score0.472
Human Vibe Score0.0922477795435268
rphilipzhangMar 25, 2025

AI-PhD-S24

Artificial Intelligence for Business Research (Spring 2024) Scribed Lecture Notes Class Recordings (You need to apply for access.) Teaching Team Instructor*: Renyu (Philip) Zhang, Associate Professor, Department of Decisions, Operations and Technology, CUHK Business School, philipzhang@cuhk.edu.hk, @911 Cheng Yu Tung Building. Teaching Assistant*: Leo Cao, Full-time TA, Department of Decisions, Operations and Technology, CUHK Business School, yinglyucao@cuhk.edu.hk. Please be noted that Leo will help with any issues related to the logistics, but not the content, of this course. Tutorial Instructor*: Qiansiqi Hu, MSBA Student, Department of Decisions, Operations and Technology, CUHK Business School, 1155208353@link.cuhk.edu.hk. BS in ECE, Shanghai Jiaotong University Michigan Institute. Basic Information Website: https://github.com/rphilipzhang/AI-PhD-S24 Time: Tuesday, 12:30pm-3:15pm, from Jan 9, 2024 to Apr 16, 2024, except for Feb 13 (Chinese New Year) and Mar 5 (Final Project Discussion) Location: Cheng Yu Tung Building (CYT) LT5 About Welcome to the mono-repo of the PhD course AI for Business Research (DSME 6635) at CUHK Business School in Spring 2024. You may download the Syllabus of this course first. The purpose of this course is to learn the following: Have a basic understanding of the fundamental concepts/methods in machine learning (ML) and artificial intelligence (AI) that are used (or potentially useful) in business research. Understand how business researchers have utilized ML/AI and what managerial questions have been addressed by ML/AI in the recent decade. Nurture a taste of what the state-of-the-art AI/ML technologies can do in the ML/AI community and, potentially, in your own research field. We will meet each Tuesday at 12:30pm in Cheng Yu Tung Building (CYT) LT5 (please pay attention to this room change). Please ask for my approval if you need to join us via the following Zoom links: Zoom link, Meeting ID 996 4239 3764, Passcode 386119. Most of the code in this course will be distributed through the Google CoLab cloud computing environment to avoid the incompatibility and version control issues on your local individual computer. On the other hand, you can always download the Jupyter Notebook from CoLab and run it your own computer. The CoLab files of this course can be found at this folder. The Google Sheet to sign up for groups and group tasks can be found here. The overleaf template for scribing the lecture notes of this course can be found here. If you have any feedback on this course, please directly contact Philip at philipzhang@cuhk.edu.hk and we will try our best to address it. Brief Schedule Subject to modifications. All classes start at 12:30pm and end at 3:15pm. |Session|Date |Topic|Key Words| |:-------:|:-------------:|:----:|:-:| |1|1.09|AI/ML in a Nutshell|Course Intro, ML Models, Model Evaluations| |2|1.16|Intro to DL|DL Intro, Neural Nets, Computational Issues in DL| |3|1.23|Prediction and Traditional NLP|Prediction in Biz Research, Pre-processing| |4|1.30|NLP (II): Traditional NLP|$N$-gram, NLP Performance Evaluations, Naïve Bayes| |5|2.06|NLP (III): Word2Vec|CBOW, Skip Gram| |6|2.20|NLP (IV): RNN|Glove, Language Model Evaluation, RNN| |7|2.27|NLP (V): Seq2Seq|LSTM, Seq2Seq, Attention Mechanism| |7.5|3.05|NLP (V.V): Transformer|The Bitter Lesson, Attention is All You Need| |8|3.12|NLP (VI): Pre-training|Computational Tricks in DL, BERT, GPT| |9|3.19|NLP (VII): LLM|Emergent Abilities, Chain-of-Thought, In-context Learning, GenAI in Business Research| |10|3.26|CV (I): Image Classification|CNN, AlexNet, ResNet, ViT| |11|4.02|CV (II): Image Segmentation and Video Analysis|R-CNN, YOLO, 3D-CNN| |12|4.09|Unsupervised Learning (I): Clustering & Topic Modeling|GMM, EM Algorithm, LDA| |13|4.16|Unsupervised Learning (II): Diffusion Models|VAE, DDPM, LDM, DiT| Important Dates All problem sets are due at 12:30pm right before class. |Date| Time|Event|Note| |:--:|:-:|:---:|:--:| |1.10| 11:59pm|Group Sign-Ups|Each group has at most two students.| |1.12| 7:00pm-9:00pm|Python Tutorial|Given by Qiansiqi Hu, Python Tutorial CoLab| |1.19| 7:00pm-9:00pm|PyTorch Tutorial|Given by Qiansiqi Hu, PyTorch Tutorial CoLab| |3.05|9:00am-6:00pm|Final Project Discussion|Please schedule a meeting with Philip.| |3.12| 12:30pm|Final Project Proposal|1-page maximum| |4.30| 11:59pm|Scribed Lecture Notes|Overleaf link| |5.12|11:59pm|Project Paper, Slides, and Code|Paper page limit: 10| Useful Resources Find more on the Syllabus. Books: ESL, Deep Learning, Dive into Deep Learning, ML Fairness, Applied Causal Inference Powered by ML and AI Courses: ML Intro by Andrew Ng, DL Intro by Andrew Ng, NLP (CS224N) by Chris Manning, CV (CS231N) by Fei-Fei Li, Deep Unsupervised Learning by Pieter Abbeel, DLR by Sergey Levine, DL Theory by Matus Telgarsky, LLM by Danqi Chen, Generative AI by Andrew Ng, Machine Learning and Big Data by Melissa Dell and Matthew Harding, Digital Economics and the Economics of AI by Martin Beraja, Chiara Farronato, Avi Goldfarb, and Catherine Tucker Detailed Schedule The following schedule is tentative and subject to changes. Session 1. Artificial Intelligence and Machine Learning in a Nutshell (Jan/09/2024) Keywords: Course Introduction, Machine Learning Basics, Bias-Variance Trade-off, Cross Validation, $k$-Nearest Neighbors, Decision Tree, Ensemble Methods Slides: Course Introduction, Machine Learning Basics CoLab Notebook Demos: k-Nearest Neighbors, Decision Tree Homework: Problem Set 1: Bias-Variance Trade-Off Online Python Tutorial: Python Tutorial CoLab, 7:00pm-9:00pm, Jan/12/2024 (Friday), given by Qiansiqi Hu, 1155208353@link.cuhk.edu.hk. Zoom Link, Meeting ID: 923 4642 4433, Pass code: 178146 References: The Elements of Statistical Learning (2nd Edition), 2009, by Trevor Hastie, Robert Tibshirani, Jerome Friedman, https://hastie.su.domains/ElemStatLearn/. Probabilistic Machine Learning: An Introduction, 2022, by Kevin Murphy, https://probml.github.io/pml-book/book1.html. Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106. Athey, Susan, and Guido W. Imbens. 2019. Machine learning methods that economists should know about. Annual Review of Economics 11: 685-725. Hofman, Jake M., et al. 2021. Integrating explanation and prediction in computational social science. Nature 595.7866: 181-188. Bastani, Hamsa, Dennis Zhang, and Heng Zhang. 2022. Applied machine learning in operations management. Innovative Technology at the Interface of Finance and Operations. Springer: 189-222. Kelly, Brian, and Dacheng Xiu. 2023. Financial machine learning, SSRN, https://ssrn.com/abstract=4501707. The Bitter Lesson, by Rich Sutton, which develops so far the most critical insight of AI: "The biggest lesson that can be read from 70 years of AI research is that general methods that leverage computation are ultimately the most effective, and by a large margin." Session 2. Introduction to Deep Learning (Jan/16/2024) Keywords: Random Forests, eXtreme Gradient Boosting Trees, Deep Learning Basics, Neural Nets Models, Computational Issues of Deep Learning Slides: Machine Learning Basics, Deep Learning Basics CoLab Notebook Demos: Random Forest, Extreme Gradient Boosting Tree, Gradient Descent, Chain Rule Presentation: By Xinyu Li and Qingyu Xu. Gu, Shihao, Brian Kelly, and Dacheng Xiu. 2020. Empirical asset pricing via machine learning. Review of Financial Studies 33: 2223-2273. Link to the paper. Homework: Problem Set 2: Implementing Neural Nets Online PyTorch Tutorial: PyTorch Tutorial CoLab, 7:00pm-9:00pm, Jan/19/2024 (Friday), given by Qiansiqi Hu, 1155208353@link.cuhk.edu.hk. Zoom Link, Meeting ID: 923 4642 4433, Pass code: 178146 References: Deep Learning, 2016, by Ian Goodfellow, Yoshua Bengio and Aaron Courville, https://www.deeplearningbook.org/. Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. Probabilistic Machine Learning: Advanced Topics, 2023, by Kevin Murphy, https://probml.github.io/pml-book/book2.html. Deep Learning with PyTorch, 2020, by Eli Stevens, Luca Antiga, and Thomas Viehmann. Gu, Shihao, Brian Kelly, and Dacheng Xiu. 2020. Empirical asset pricing with machine learning. Review of Financial Studies 33: 2223-2273. Session 3. DL Basics, Predictions in Business Research, and Traditonal NLP (Jan/23/2024) Keywords: Optimization and Computational Issues of Deep Learning, Prediction Problems in Business Research, Pre-processing and Word Representations in Traditional Natural Language Processing Slides: Deep Learning Basics, Prediction Problems in Business Research, NLP(I): Pre-processing and Word Representations.pdf) CoLab Notebook Demos: He Initialization, Dropout, Micrograd, NLP Pre-processing Presentation: By Letian Kong and Liheng Tan. Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106. Link to the paper. Homework: Problem Set 2: Implementing Neural Nets, due at 12:30pm, Jan/30/2024 (Tuesday). References: Kleinberg, Jon, Jens Ludwig, Sendhil Mullainathan, and Ziad Obermeyer. 2015. Prediction policy problems. American Economic Review 105(5): 491-495. Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106. Kleinberg, Jon, Himabindu Lakkaraju, Jure Leskovec, Jens Ludwig, and Sendhil Mullainathan. 2018. Human decisions and machine predictions. Quarterly Journal of Economics 133(1): 237-293. Bajari, Patrick, Denis Nekipelov, Stephen P. Ryan, and Miaoyu Yang. 2015. Machine learning methods for demand estimation. American Economic Review, 105(5): 481-485. Farias, Vivek F., and Andrew A. Li. 2019. Learning preferences with side information. Management Science 65(7): 3131-3149. Cui, Ruomeng, Santiago Gallino, Antonio Moreno, and Dennis J. Zhang. 2018. The operational value of social media information. Production and Operations Management, 27(10): 1749-1769. Gentzkow, Matthew, Bryan Kelly, and Matt Taddy. 2019. Text as data. Journal of Economic Literature, 57(3): 535-574. Chapter 2, Introduction to Information Retrieval, 2008, Cambridge University Press, by Christopher D. Manning, Prabhakar Raghavan and Hinrich Schutze, https://nlp.stanford.edu/IR-book/information-retrieval-book.html. Chapter 2, Speech and Language Processing (3rd ed. draft), 2023, by Dan Jurafsky and James H. Martin, https://web.stanford.edu/~jurafsky/slp3/. Parameter Initialization and Batch Normalization (in Chinese) GPU Comparisons-vs-NVIDIA-H100-(PCIe)-vs-NVIDIA-RTX-6000-Ada/624vs632vs640) GitHub Repo for Micrograd, by Andrej Karpathy. Hand Written Notes Session 4. Traditonal NLP (Jan/30/2024) Keywords: Pre-processing and Word Representations in NLP, N-Gram, Naïve Bayes, Language Model Evaluation, Traditional NLP Applied to Business/Econ Research Slides: NLP(I): Pre-processing and Word Representations.pdf), NLP(II): N-Gram, Naïve Bayes, and Language Model Evaluation.pdf) CoLab Notebook Demos: NLP Pre-processing, N-Gram, Naïve Bayes Presentation: By Zhi Li and Boya Peng. Hansen, Stephen, Michael McMahon, and Andrea Prat. 2018. Transparency and deliberation within the FOMC: A computational linguistics approach. Quarterly Journal of Economics, 133(2): 801-870. Link to the paper. Homework: Problem Set 3: Implementing Traditional NLP Techniques, due at 12:30pm, Feb/6/2024 (Tuesday). References: Gentzkow, Matthew, Bryan Kelly, and Matt Taddy. 2019. Text as data. Journal of Economic Literature, 57(3): 535-574. Hansen, Stephen, Michael McMahon, and Andrea Prat. 2018. Transparency and deliberation within the FOMC: A computational linguistics approach. Quarterly Journal of Economics, 133(2): 801-870. Chapters 2, 12, & 13, Introduction to Information Retrieval, 2008, Cambridge University Press, by Christopher D. Manning, Prabhakar Raghavan and Hinrich Schutze, https://nlp.stanford.edu/IR-book/information-retrieval-book.html. Chapter 2, 3 & 4, Speech and Language Processing (3rd ed. draft), 2023, by Dan Jurafsky and James H. Martin, https://web.stanford.edu/~jurafsky/slp3/. Natural Language Tool Kit (NLTK) Documentation Hand Written Notes Session 5. Deep-Learning-Based NLP: Word2Vec (Feb/06/2024) Keywords: Traditional NLP Applied to Business/Econ Research, Word2Vec: Continuous Bag of Words and Skip-Gram Slides: NLP(II): N-Gram, Naïve Bayes, and Language Model Evaluation.pdf), NLP(III): Word2Vec.pdf) CoLab Notebook Demos: Word2Vec: CBOW, Word2Vec: Skip-Gram Presentation: By Xinyu Xu and Shu Zhang. Timoshenko, Artem, and John R. Hauser. 2019. Identifying customer needs from user-generated content. Marketing Science, 38(1): 1-20. Link to the paper. Homework: No homework this week. Probably you should think about your final project when enjoying your Lunar New Year Holiday. References: Gentzkow, Matthew, Bryan Kelly, and Matt Taddy. 2019. Text as data. Journal of Economic Literature, 57(3): 535-574. Tetlock, Paul. 2007. Giving content to investor sentiment: The role of media in the stock market. Journal of Finance, 62(3): 1139-1168. Baker, Scott, Nicholas Bloom, and Steven Davis, 2016. Measuring economic policy uncertainty. Quarterly Journal of Economics, 131(4): 1593-1636. Gentzkow, Matthew, and Jesse Shapiro. 2010. What drives media slant? Evidence from US daily newspapers. Econometrica, 78(1): 35-71. Timoshenko, Artem, and John R. Hauser. 2019. Identifying customer needs from user-generated content. Marketing Science, 38(1): 1-20. Mikolov, Tomas, Kai Chen, Greg Corrado, and Jeff Dean. 2013. Efficient estimation of word representations in vector space. ArXiv Preprint, arXiv:1301.3781. Mikolov, Tomas, Ilya Sutskever, Kai Chen, Greg Corrado, and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. Advances in Neural Information Processing Systems (NeurIPS) 26. Parts I - II, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Word Embeddings Trained on Google News Corpus Hand Written Notes Session 6. Deep-Learning-Based NLP: RNN and Seq2Seq (Feb/20/2024) Keywords: Word2Vec: GloVe, Word Embedding and Language Model Evaluations, Word2Vec and RNN Applied to Business/Econ Research, RNN Slides: Guest Lecture Announcement, NLP(III): Word2Vec.pdf), NLP(IV): RNN & Seq2Seq.pdf) CoLab Notebook Demos: Word2Vec: CBOW, Word2Vec: Skip-Gram Presentation: By Qiyu Dai and Yifan Ren. Huang, Allen H., Hui Wang, and Yi Yang. 2023. FinBERT: A large language model for extracting information from financial text. Contemporary Accounting Research, 40(2): 806-841. Link to the paper. Link to GitHub Repo. Homework: Problem Set 4 - Word2Vec & LSTM for Sentiment Analysis References: Ash, Elliot, and Stephen Hansen. 2023. Text algorithms in economics. Annual Review of Economics, 15: 659-688. Associated GitHub with Code Demonstrations. Li, Kai, Feng Mai, Rui Shen, and Xinyan Yan. 2021. Measuring corporate culture using machine learning. Review of Financial Studies, 34(7): 3265-3315. Chen, Fanglin, Xiao Liu, Davide Proserpio, and Isamar Troncoso. 2022. Product2Vec: Leveraging representation learning to model consumer product choice in large assortments. Available at SSRN 3519358. Pennington, Jeffrey, Richard Socher, and Christopher Manning. 2014. Glove: Global vectors for word representation. Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP) (pp. 1532-1543). Parts 2 and 5, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Chapters 9 and 10, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. RNN and LSTM Visualizations Hand Written Notes Session 7. Deep-Learning-Based NLP: Attention and Transformer (Feb/27/2024) Keywords: RNN and its Applications to Business/Econ Research, LSTM, Seq2Seq, Attention Mechanism Slides: Final Project, NLP(IV): RNN & Seq2Seq.pdf), NLP(V): Attention & Transformer.pdf) CoLab Notebook Demos: RNN & LSTM, Attention Mechanism Presentation: By Qinghe Gui and Chaoyuan Jiang. Zhang, Mengxia and Lan Luo. 2023. Can consumer-posted photos serve as a leading indicator of restaurant survival? Evidence from Yelp. Management Science 69(1): 25-50. Link to the paper. Homework: Problem Set 4 - Word2Vec & LSTM for Sentiment Analysis References: Qi, Meng, Yuanyuan Shi, Yongzhi Qi, Chenxin Ma, Rong Yuan, Di Wu, Zuo-Jun (Max) Shen. 2023. A Practical End-to-End Inventory Management Model with Deep Learning. Management Science, 69(2): 759-773. Sarzynska-Wawer, Justyna, Aleksander Wawer, Aleksandra Pawlak, Julia Szymanowska, Izabela Stefaniak, Michal Jarkiewicz, and Lukasz Okruszek. 2021. Detecting formal thought disorder by deep contextualized word representations. Psychiatry Research, 304, 114135. Hansen, Stephen, Peter J. Lambert, Nicholas Bloom, Steven J. Davis, Raffaella Sadun, and Bledi Taska. 2023. Remote work across jobs, companies, and space (No. w31007). National Bureau of Economic Research. Sutskever, Ilya, Oriol Vinyals, and Quoc V. Le. 2014. Sequence to sequence learning with neural networks. Advances in neural information processing systems, 27. Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. ICLR Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... and Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems, 30. Parts 5, 6, and 8, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Chapters 9, 10, and 11, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. RNN and LSTM Visualizations PyTorch's Tutorial of Seq2Seq for Machine Translation Illustrated Transformer Transformer from Scratch, with the Code on GitHub Hand Written Notes Session 7.5. Deep-Learning-Based NLP: Attention is All You Need (Mar/05/2024) Keywords: Bitter Lesson: Power of Computation in AI, Attention Mechanism, Transformer Slides: The Bitter Lesson, NLP(V): Attention & Transformer.pdf) CoLab Notebook Demos: Attention Mechanism, Transformer Homework: One-page Proposal for Your Final Project References: The Bitter Lesson, by Rich Sutton Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. ICLR Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... and Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems, 30. Part 8, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Chapter 11, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. Illustrated Transformer Transformer from Scratch, with the Code on GitHub Andrej Karpathy's Lecture to Build Transformers Hand Written Notes Session 8. Deep-Learning-Based NLP: Pretraining (Mar/12/2024) Keywords: Computations in AI, BERT (Bidirectional Encoder Representations from Transformers), GPT (Generative Pretrained Transformers) Slides: Guest Lecture by Dr. Liubo Li on Deep Learning Computation, Pretraining.pdf) CoLab Notebook Demos: Crafting Intelligence: The Art of Deep Learning Modeling, BERT API @ Hugging Face Presentation: By Zhankun Chen and Yiyi Zhao. Noy, Shakked and Whitney Zhang. 2023. Experimental evidence on the productivity effects of generative artificial intelligence. Science, 381: 187-192. Link to the Paper Homework: Problem Set 5 - Sentiment Analysis with Hugging Face, due at 12:30pm, March 26, Tuesday. References: Devlin, Jacob, Ming-Wei Chang, Kenton Lee, Kristina Toutanova. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. ArXiv preprint arXiv:1810.04805. GitHub Repo Radford, Alec, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. 2018. Improving language understanding by generative pre-training, (GPT-1) PDF link, GitHub Repo Radford, Alec, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI blog, 1(8), 9. (GPT-2) PDF Link, GitHub Repo Brown, Tom, et al. 2020. Language models are few-shot learners. Advances in neural information processing systems, 33, 1877-1901. (GPT-3) GitHub Repo Huang, Allen H., Hui Wang, and Yi Yang. 2023. FinBERT: A large language model for extracting information from financial text. Contemporary Accounting Research, 40(2): 806-841. GitHub Repo Part 9, Lecture Notes and Slides for CS 224N: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto. Link to CS 224N Part 2 & 4, Slides for COS 597G: Understanding Large Language Models, by Danqi Chen. Link to COS 597G A Visual Guide to BERT, How GPT-3 Works Andrej Karpathy's Lecture to Build GPT-2 (124M) from Scratch Hand Written Notes Session 9. Deep-Learning-Based NLP: Large Language Models (Mar/19/2024) Keywords: Large Language Models, Generative AI, Emergent Ababilities, Instruction Fine-Tuning (IFT), Reinforcement Learning with Human Feedback (RLHF), In-Context Learning, Chain-of-Thought (CoT) Slides: What's Next, Pretraining.pdf), Large Language Models.pdf) CoLab Notebook Demos: BERT API @ Hugging Face Presentation: By Jia Liu. Liu, Liu, Dzyabura, Daria, Mizik, Natalie. 2020. Visual listening in: Extracting brand image portrayed on social media. Marketing Science, 39(4): 669-686. Link to the Paper Homework: Problem Set 5 - Sentiment Analysis with Hugging Face, due at 12:30pm, March 26, Tuesday (soft-deadline). References: Wei, Jason, et al. 2021. Finetuned language models are zero-shot learners. ArXiv preprint arXiv:2109.01652, link to the paper. Wei, Jason, et al. 2022. Emergent abilities of large language models. ArXiv preprint arXiv:2206.07682, link to the paper. Ouyang, Long, et al. 2022. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35, 27730-27744. Wei, Jason, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35, 24824-24837. Kaplan, Jared. 2020. Scaling laws for neural language models. ArXiv preprint arXiv:2001.08361, link to the paper. Hoffmann, Jordan, et al. 2022. Training compute-optimal large language models. ArXiv preprint arXiv:2203.15556, link to the paper. Shinn, Noah, et al. 2023. Reflexion: Language agents with verbal reinforcement learning. ArXiv preprint arXiv:2303.11366, link to the paper. Reisenbichler, Martin, Thomas Reutterer, David A. Schweidel, and Daniel Dan. 2022. Frontiers: Supporting content marketing with natural language generation. Marketing Science, 41(3): 441-452. Romera-Paredes, B., Barekatain, M., Novikov, A. et al. 2023. Mathematical discoveries from program search with large language models. Nature, link to the paper. Part 10, Lecture Notes and Slides for CS224N: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto. Link to CS 224N COS 597G: Understanding Large Language Models, by Danqi Chen. Link to COS 597G Andrej Karpathy's 1-hour Talk on LLM CS224n, Hugging Face Tutorial Session 10. Deep-Learning-Based CV: Image Classification (Mar/26/2024) Keywords: Large Language Models Applications, Convolution Neural Nets (CNN), LeNet, AlexNet, VGG, ResNet, ViT Slides: What's Next, Large Language Models.pdf), Image Classification.pdf) CoLab Notebook Demos: CNN, LeNet, & AlexNet, VGG, ResNet, ViT Presentation: By Yingxin Lin and Zeshen Ye. Netzer, Oded, Alain Lemaire, and Michal Herzenstein. 2019. When words sweat: Identifying signals for loan default in the text of loan applications. Journal of Marketing Research, 56(6): 960-980. Link to the Paper Homework: Problem Set 6 - AlexNet and ResNet, due at 12:30pm, April 9, Tuesday. References: Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton. 2012. Imagenet classification with deep convolutional neural networks. Advances in Neural Information Processing Systems, 25. He, Kaiming, Xiangyu Zhang, Shaoqing Ren and Jian Sun. 2016. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition, 770-778. Dosovitskiy, Alexey, et al. 2020. An image is worth 16x16 words: Transformers for image recognition at scale. ArXiv preprint, arXiv:2010.11929, link to the paper, link to the GitHub repo. Jean, Neal, Marshall Burke, Michael Xie, Matthew W. Davis, David B. Lobell, and Stefand Ermon. 2016. Combining satellite imagery and machine learning to predict poverty. Science, 353(6301), 790-794. Zhang, Mengxia and Lan Luo. 2023. Can consumer-posted photos serve as a leading indicator of restaurant survival? Evidence from Yelp. Management Science 69(1): 25-50. Course Notes (Lectures 5 & 6) for CS231n: Deep Learning for Computer Vision, by Fei-Fei Li, Ruohan Gao, & Yunzhu Li. Link to CS231n. Chapters 7 and 8, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola. Link to the book. Fine-Tune ViT for Image Classification with Hugging Face 🤗 Transformers Hugging Face 🤗 ViT CoLab Tutorial Session 11. Deep-Learning-Based CV (II): Object Detection & Video Analysis (Apr/2/2024) Keywords: Image Processing Applications, Localization, R-CNNs, YOLOs, Semantic Segmentation, 3D CNN, Video Analysis Applications Slides: What's Next, Image Classification.pdf), Object Detection and Video Analysis.pdf) CoLab Notebook Demos: Data Augmentation, Faster R-CNN & YOLO v5 Presentation: By Qinlu Hu and Yilin Shi. Yang, Jeremy, Juanjuan Zhang, and Yuhan Zhang. 2023. Engagement that sells: Influencer video advertising on TikTok. Available at SSRN Link to the Paper Homework: Problem Set 6 - AlexNet and ResNet, due at 12:30pm, April 9, Tuesday. References: Girshick, R., Donahue, J., Darrell, T. and Malik, J., 2014. Rich feature hierarchies for accurate object detection and semantic segmentation. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 580-587). Redmon, Joseph, Santosh Divvala, Ross Girshick, and Ali Farhadi. 2016. You only look once: Unified, real-time object detection. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 779-788). Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R. and Fei-Fei, L., 2014. Large-scale video classification with convolutional neural networks. Proceedings of the IEEE conference on Computer Vision and Pattern Recognition (pp. 1725-1732). Glaeser, Edward L., Scott D. Kominers, Michael Luca, and Nikhil Naik. 2018. Big data and big cities: The promises and limitations of improved measures of urban life. Economic Inquiry, 56(1): 114-137. Zhang, S., Xu, K. and Srinivasan, K., 2023. Frontiers: Unmasking Social Compliance Behavior During the Pandemic. Marketing Science, 42(3), pp.440-450. Course Notes (Lectures 10 & 11) for CS231n: Deep Learning for Computer Vision, by Fei-Fei Li, Ruohan Gao, & Yunzhu Li. Link to CS231n. Chapter 14, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola. Link to the book. Hand Written Notes Session 12. Unsupervised Learning: Clustering, Topic Modeling & VAE (Apr/9/2024) Keywords: K-Means, Gaussian Mixture Models, EM-Algorithm, Latent Dirichlet Allocation, Variational Auto-Encoder Slides: What's Next, Clustering, Topic Modeling & VAE.pdf) CoLab Notebook Demos: K-Means, LDA, VAE Homework: Problem Set 7 - Unsupervised Learning (EM & LDA), due at 12:30pm, April 23, Tuesday. References: Blei, David M., Ng, Andrew Y., and Jordan, Michael I. 2003. Latent Dirichlet allocation. Journal of Machine Learning Research, 3(Jan): 993-1022. Kingma, D.P. and Welling, M., 2013. Auto-encoding Variational Bayes. arXiv preprint arXiv:1312.6114. Kingma, D.P. and Welling, M., 2019. An introduction to variational autoencoders. Foundations and Trends® in Machine Learning, 12(4), pp.307-392. Bandiera, O., Prat, A., Hansen, S., & Sadun, R. 2020. CEO behavior and firm performance. Journal of Political Economy, 128(4), 1325-1369. Liu, Jia and Olivier Toubia. 2018. A semantic approach for estimating consumer content preferences from online search queries. Marketing Science, 37(6): 930-952. Mueller, Hannes, and Christopher Rauh. 2018. Reading between the lines: Prediction of political violence using newspaper text. American Political Science Review, 112(2): 358-375. Tian, Z., Dew, R. and Iyengar, R., 2023. Mega or Micro? Influencer Selection Using Follower Elasticity. Journal of Marketing Research. Chapters 8.5 and 14, The Elements of Statistical Learning (2nd Edition), 2009, by Trevor Hastie, Robert Tibshirani, Jerome Friedman, Link to Book. Course Notes (Lectures 1 & 4) for CS294-158-SP24: Deep Unsupervised Learning, taught by Pieter Abbeel, Wilson Yan, Kevin Frans, Philipp Wu. Link to CS294-158-SP24. Hand Written Notes Session 13. Unsupervised Learning: Diffusion Models (Apr/16/2024) Keywords: VAE, Denoised Diffusion Probabilistic Models, Latent Diffusion Models, CLIP, Imagen, Diffusion Transformers Slides: Clustering, Topic Modeling & VAE.pdf), Diffusion Models.pdf), Course Summary CoLab Notebook Demos: VAE, DDPM, DiT Homework: Problem Set 7 - Unsupervised Learning (EM & LDA), due at 12:30pm, April 23, Tuesday. References: Kingma, D.P. and Welling, M., 2013. Auto-encoding Variational Bayes. arXiv preprint arXiv:1312.6114. Kingma, D.P. and Welling, M., 2019. An introduction to variational autoencoders. Foundations and Trends® in Machine Learning, 12(4), pp.307-392. Ho, J., Jain, A. and Abbeel, P., 2020. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33, 6840-6851. Chan, S.H., 2024. Tutorial on Diffusion Models for Imaging and Vision. arXiv preprint arXiv:2403.18103. Peebles, W. and Xie, S., 2023. Scalable diffusion models with transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 4195-4205. Link to GitHub Repo. Tian, Z., Dew, R. and Iyengar, R., 2023. Mega or Micro? Influencer Selection Using Follower Elasticity. Journal of Marketing Research. Ludwig, J. and Mullainathan, S., 2024. Machine learning as a tool for hypothesis generation. Quarterly Journal of Economics, 139(2), 751-827. Burnap, A., Hauser, J.R. and Timoshenko, A., 2023. Product aesthetic design: A machine learning augmentation. Marketing Science, 42(6), 1029-1056. Course Notes (Lecture 6) for CS294-158-SP24: Deep Unsupervised Learning, taught by Pieter Abbeel, Wilson Yan, Kevin Frans, Philipp Wu. Link to CS294-158-SP24. CVPR 2022 Tutorial: Denoising Diffusion-based Generative Modeling: Foundations and Applications, by Karsten Kreis, Ruiqi Gao, and Arash Vahdat Link to the Tutorial Lilian Weng (OpenAI)'s Blog on Diffusion Models Lilian Weng (OpenAI)'s Blog on Diffusion Models for Video Generation Hugging Face Diffusers 🤗 Library Hand Written Notes

voicefilter
github
LLM Vibe Score0.496
Human Vibe Score0.029786815978503328
maum-aiMar 24, 2025

voicefilter

VoiceFilter Note from Seung-won (2020.10.25) Hi everyone! It's Seung-won from MINDs Lab, Inc. It's been a long time since I've released this open-source, and I didn't expect this repository to grab such a great amount of attention for a long time. I would like to thank everyone for giving such attention, and also Mr. Quan Wang (the first author of the VoiceFilter paper) for referring this project in his paper. Actually, this project was done by me when it was only 3 months after I started studying deep learning & speech separation without a supervisor in the relevant field. Back then, I didn't know what is a power-law compression, and the correct way to validate/test the models. Now that I've spent more time on deep learning & speech since then (I also wrote a paper published at Interspeech 2020 😊), I can observe some obvious mistakes that I've made. Those issues were kindly raised by GitHub users; please refer to the Issues and Pull Requests for that. That being said, this repository can be quite unreliable, and I would like to remind everyone to use this code at their own risk (as specified in LICENSE). Unfortunately, I can't afford extra time on revising this project or reviewing the Issues / Pull Requests. Instead, I would like to offer some pointers to newer, more reliable resources: VoiceFilter-Lite: This is a newer version of VoiceFilter presented at Interspeech 2020, which is also written by Mr. Quan Wang (and his colleagues at Google). I highly recommend checking this paper, since it focused on a more realistic situation where VoiceFilter is needed. List of VoiceFilter implementation available on GitHub: In March 2019, this repository was the only available open-source implementation of VoiceFilter. However, much better implementations that deserve more attention became available across GitHub. Please check them, and choose the one that meets your demand. PyTorch Lightning: Back in 2019, I could not find a great deep-learning project template for myself, so I and my colleagues had used this project as a template for other new projects. For people who are searching for such project template, I would like to strongly recommend PyTorch Lightning. Even though I had done a lot of effort into developing my own template during 2019 (VoiceFilter -> RandWireNN -> MelNet -> MelGAN), I found PyTorch Lightning much better than my own template. Thanks for reading, and I wish everyone good health during the global pandemic situation. Best regards, Seung-won Park Unofficial PyTorch implementation of Google AI's: VoiceFilter: Targeted Voice Separation by Speaker-Conditioned Spectrogram Masking. Result Training took about 20 hours on AWS p3.2xlarge(NVIDIA V100). Audio Sample Listen to audio sample at webpage: http://swpark.me/voicefilter/ Metric | Median SDR | Paper | Ours | | ---------------------- | ----- | ---- | | before VoiceFilter | 2.5 | 1.9 | | after VoiceFilter | 12.6 | 10.2 | SDR converged at 10, which is slightly lower than paper's. Dependencies Python and packages This code was tested on Python 3.6 with PyTorch 1.0.1. Other packages can be installed by: Miscellaneous ffmpeg-normalize is used for resampling and normalizing wav files. See README.md of ffmpeg-normalize for installation. Prepare Dataset Download LibriSpeech dataset To replicate VoiceFilter paper, get LibriSpeech dataset at http://www.openslr.org/12/. train-clear-100.tar.gz(6.3G) contains speech of 252 speakers, and train-clear-360.tar.gz(23G) contains 922 speakers. You may use either, but the more speakers you have in dataset, the more better VoiceFilter will be. Resample & Normalize wav files First, unzip tar.gz file to desired folder: Next, copy utils/normalize-resample.sh to root directory of unzipped data folder. Then: Edit config.yaml Preprocess wav files In order to boost training speed, perform STFT for each files before training by: This will create 100,000(train) + 1000(test) data. (About 160G) Train VoiceFilter Get pretrained model for speaker recognition system VoiceFilter utilizes speaker recognition system (d-vector embeddings). Here, we provide pretrained model for obtaining d-vector embeddings. This model was trained with VoxCeleb2 dataset, where utterances are randomly fit to time length [70, 90] frames. Tests are done with window 80 / hop 40 and have shown equal error rate about 1%. Data used for test were selected from first 8 speakers of VoxCeleb1 test dataset, where 10 utterances per each speakers are randomly selected. Update: Evaluation on VoxCeleb1 selected pair showed 7.4% EER. The model can be downloaded at this GDrive link. Run After specifying traindir, testdir at config.yaml, run: This will create chkpt/name and logs/name at base directory(-b option, . in default) View tensorboardX Resuming from checkpoint Evaluate Possible improvments Try power-law compressed reconstruction error as loss function, instead of MSE. (See #14) Author Seungwon Park at MINDsLab (yyyyy@snu.ac.kr, swpark@mindslab.ai) License Apache License 2.0 This repository contains codes adapted/copied from the followings: utils/adabound.py from https://github.com/Luolc/AdaBound (Apache License 2.0) utils/audio.py from https://github.com/keithito/tacotron (MIT License) utils/hparams.py from https://github.com/HarryVolek/PyTorchSpeakerVerification (No License specified) utils/normalize-resample.sh from https://unix.stackexchange.com/a/216475

video-killed-the-radio-star
github
LLM Vibe Score0.48
Human Vibe Score0.018384486870142776
dmarxMar 23, 2025

video-killed-the-radio-star

Video Killed The Radio Star Requirements ffmpeg - https://ffmpeg.org/ pytorch - https://pytorch.org/get-started/locally/ vktrs - (this repo) - pip install vktrs[api] stability_sdk api token - https://beta.dreamstudio.ai/ > circular icon in top right > membership > API Key whisper - pip install git+https://github.com/openai/whisper FAQ What is this? TLDR: Automated music video maker, given an mp3 or a youtube URL How does this animation technique work? For each text prompt you provide, the notebook will... Generate an image based on that text prompt (using stable diffusion) Use the generated image as the init_image to recombine with the text prompt to generate variations similar to the first image. This produces a sequence of extremely similar images based on the original text prompt Images are then intelligently reordered to find the smoothest animation sequence of those frames This image sequence is then repeated to pad out the animation duration as needed The technique demonstrated in this notebook was inspired by a video created by Ben Gillin. How are lyrics transcribed? This notebook uses openai's recently released 'whisper' model for performing automatic speech recognition. OpenAI was kind of to offer several different sizes of this model which each have their own pros and cons. This notebook uses the largest whisper model for transcribing the actual lyrics. Additionally, we use the smallest model for performing the lyric segmentation. Neither of these models is perfect, but the results so far seem pretty decent. The first draft of this notebook relied on subtitles from youtube videos to determine timing, which was then aligned with user-provided lyrics. Youtube's automated captions are powerful and I'll update the notebook shortly to leverage those again, but for the time being we're just using whisper for everything and not referencing user-provided captions at all. Something didn't work quite right in the transcription process. How do fix the timing or the actual lyrics? The notebook is divided into several steps. Between each step, a "storyboard" file is updated. If you want to make modifications, you can edit this file directly and those edits should be reflected when you next load the file. Depending on what you changed and what step you run next, your changes may be ignored or even overwritten. Still playing with different solutions here. Can I provide my own images to 'bring to life' and associate with certain lyrics/sequences? Yes, you can! As described above: you just need to modify the storyboard. Will describe this functionality in greater detail after the implementation stabilizes a bit more. This gave me an idea and I'd like to use just a part of your process here. What's the best way to reuse just some of the machinery you've developed here? Most of the functionality in this notebook has been offloaded to library I published to pypi called vktrs. I strongly encourage you to import anything you need from there rather than cutting and pasting function into a notebook. Similarly, if you have ideas for improvements, please don't hesitate to submit a PR! Dev notes

How-to-learn-Deep-Learning
github
LLM Vibe Score0.524
Human Vibe Score0.1392403398579415
emilwallnerMar 23, 2025

How-to-learn-Deep-Learning

Approach A practical, top-down approach, starting with high-level frameworks with a focus on Deep Learning. UPDATED VERSION: 👉 Check out my 60-page guide, No ML Degree, on how to land a machine learning job without a degree. Getting started [2 months] There are three main goals to get up to speed with deep learning: 1) Get familiar to the tools you will be working with, e.g. Python, the command line and Jupyter notebooks 2) Get used to the workflow, everything from finding the data to deploying a trained model 3) Building a deep learning mindset, an intuition for how deep learning models behave and how to improve them Spend a week on codecademy.com and learn the python syntax, command line and git. If you don't have any previous programming experience, it's good to spend a few months learning how to program. Otherwise, it's easy to become overwhelmed. Spend one to two weeks using Pandas and Scikit-learn on Kaggle problems using Jupyter Notebook on Colab, e.g. Titanic, House prices, and Iris. This gives you an overview of the machine learning mindset and workflow. Spend one month implementing models on cloud GPUs. Start with FastAI and PyTorch. The FastAI community is the go-to place for people wanting to apply deep learning and share the state of the art techniques. Once you have done this, you will know how to add value with ML. Portfolio [3 - 12 months] Think of your portfolio as evidence to a potential employer that you can provide value for them. When you are looking for your first job, there are four main roles you can apply for Machine Learning Engineering, Applied Machine Learning Researcher / Residencies, Machine Learning Research Scientist, and Software Engineering. A lot of the work related to machine learning is pure software engineering roles (category 4), e.g. scaling infrastructure, but that's out of scope for this article. It's easiest to get a foot in the door if you aim for Machine Learning Engineering roles. There are a magnitude more ML engineering roles compared to category 2 & 3 roles, they require little to no theory, and they are less competitive. Most employers prefer scaling and leveraging stable implementations, often ~1 year old, instead of allocating scarce resources to implement SOTA papers, which are often time-consuming and seldom work well in practice. Once you can cover your bills and have a few years of experience, you are in a better position to learn theory and advance to category 2 & 3 roles. This is especially true if you are self-taught, you often have an edge against an average university graduate. In general, graduates have weak practical skills and strong theory skills. Context You'll have a mix of 3 - 10 technical and non-technical people looking at your portfolio, regardless of their background, you want to spark the following reactions: the applicant has experience tackling our type of problems, the applicant's work is easy to understand and well organized, and the work was without a doubt 100% made by the applicant. Most ML learners end up with the same portfolio as everyone else. Portfolio items include things as MOOC participation, dog/cat classifiers, and implementations on toy datasets such as the titanic and iris datasets. They often indicate that you actively avoid real-world problem-solving, and prefer being in your comfort zone by copy-pasting from tutorials. These portfolio items often signal negative value instead of signaling that you are a high-quality candidate. A unique portfolio item implies that you have tackled a unique problem without a solution, and thus have to engage in the type of problem-solving an employee does daily. A good starting point is to look for portfolio ideas on active Kaggle competitions, and machine learning consulting projects, and demo versions of common production pipelines. Here's a Twitter thread on how to come up with portfolio ideas. Here are rough guidelines to self-assess the strength of your portfolio: Machine learning engineering: Even though ML engineering roles are the most strategic entry point, they are still highly competitive. In general, there are ~50 software engineering roles for every ML role. From the self-learners I know, 2/3 fail to get a foot in the door and end up taking software engineering roles instead. You are ready to look for a job when you have two high-quality projects that are well-documented, have unique datasets, and are relevant to a specific industry, say banking or insurance. Project Type | Base score | -------------| -----------| Common project | -1 p || Unique project | 10 p | Multiplier Type | Factor -----------------|----------------- Strong documentation | 5x 5000-word article | 5x Kaggle Medal | 10x Employer relevancy | 20x Hireable: 5,250 p Competative: 15,000 p Applied research / research assistant/ residencies: For most companies, the risk of pursuing cutting edge research is often too high, thus only the biggest companies tend to need this skillset. There are smaller research organizations that hire for these positions, but these positions tend to be poorly advertised and have a bias for people in their existing community. Many of these roles don't require a Ph.D., which makes them available to most people with a Bachelor's or Master's degrees, or self-learners with one year of focussed study. Given the status, scarcity, and requirements for these positions, they are the most competitive ML positions. Positions at well-known companies tend to get more than a thousand applicants per position. Daily, these roles require that you understand and can implement SOTA papers, thus that's what they will be looking for in your portfolio. Projects type | Base score --------------| ----------- Common project | -10 p Unique project | 1 p SOTA paper implementation | 20 p Multiplier type | Factor ----------------| --------------- Strong documentation | 5x 5000-word article | 5x SOTA performance | 5x Employer relevancy | 20x Hireable: 52,500 p Competitive: 150,000 p Research Scientist: Research scientist roles require a Ph.D. or equivalent experience. While the former category requires the ability to implement SOTA papers, this category requires you to come up with research ideas. The mainstream research community measure the quality of research ideas by their impact, here is a list of the venues and their impact. To have a competitive portfolio, you need two published papers in the top venues in an area that's relevant to your potential employer. Project type | Base score -------------| ---------------- Common project | -100 p An unpublished paper | 5 p ICML/ICLR/NeurIPS publication | 500p All other publications | 50 p Multiplier type | Factor ------------------| ------------------ First author paper | 10x Employer relevancy | 20x Hireable: 20,000 p Competitive roles and elite PhD positions: 200,000 p Examples: My first portfolio item (after 2 months of learning): Code | Write-up My second portfolio item (after 4 months of learning): Code | Write-up Dylan Djian's first portfolio item: Code | Write-up Dylan Djian's second portfolio item: Code | Write-up Reiichiro Nakano's first portfolio item: Code | Write-up Reiichiro Nakano's second portfolio item: Write-up Most recruiters will spend 10-20 seconds on each of your portfolio items. Unless they can understand the value in that time frame, the value of the project is close to zero. Thus, writing and documentation are key. Here's another thread on how to write about portfolio items. The last key point is relevancy. It's more fun to make a wide range of projects, but if you want to optimize for breaking into the industry, you want to do all projects in one niche, thus making your skillset super relevant for a specific pool of employers. Further Inspiration: FastAI student projects Stanford NLP student projects Stanford CNN student projects Theory 101 [4 months] Learning how to read papers is critical if you want to get into research, and a brilliant asset as an ML engineer. There are three key areas to feel comfortable reading papers: 1) Understanding the details of the most frequent algorithms, gradient descent, linear regression, and MLPs, etc 2) Learning how to translate the most frequent math notations into code 3) Learn the basics of algebra, calculus, statistics, and machine learning For the first week, spend it on 3Blue1Brown's Essence of linear algebra, the Essence of Calculus, and StatQuests' the Basics (of statistics) and Machine Learning. Use a spaced repetition app like Anki and memorize all the key concepts. Use images as much as possible, they are easier to memorize. Spend one month recoding the core concepts in python numpy, including least squares, gradient descent, linear regression, and a vanilla neural network. This will help you reduce a lot of cognitive load down the line. Learning that notations are compact logic and how to translate it into code will make you feel less anxious about the theory. I believe the best deep learning theory curriculum is the Deep Learning Book by Ian Goodfellow and Yoshua Bengio and Aaron Courville. I use it as a curriculum, and the use online courses and internet resources to learn the details about each concept. Spend three months on part 1 of the Deep learning book. Use lectures and videos to understand the concepts, Khan academy type exercises to master each concept, and Anki flashcards to remember them long-term. Key Books: Deep Learning Book by Ian Goodfellow and Yoshua Bengio and Aaron Courville. Deep Learning for Coders with fastai and PyTorch: AI Applications Without a PhD by Jeremy Howard and Sylvain. Gugger. Deep Learning with Python by François Chollet. Neural Networks and Deep Learning by Michael Nielsen. Grokking Deep Learning by Andrew W. Trask. Forums FastAI Keras Slack Distill Slack Pytorch Twitter Other good learning strategies: Emil Wallner S. Zayd Enam Catherine Olsson Greg Brockman V2 Greg Brockman V1 Andrew Ng Amid Fish Spinning Up by OpenAI Confession as an AI researcher YC Threads: One and Two If you have suggestions/questions create an issue or ping me on Twitter. UPDATED VERSION: 👉 Check out my 60-page guide, No ML Degree, on how to land a machine learning job without a degree. Language versions: Korean | English

Overmind
github
LLM Vibe Score0.469
Human Vibe Score0.20474237922306593
bencbartlettMar 23, 2025

Overmind

[](https://github.com/bencbartlett/Overmind/releases) [](https://github.com/bencbartlett/Overmind/blob/master/CHANGELOG.md) [](https://bencbartlett.github.io/overmind-docs/) [](https://github.com/bencbartlett/Overmind/wiki) [](https://screeps.slack.com/messages/overmind) [](https://github.com/bencbartlett/Overmind/issues/new) [](https://github.com/bencbartlett/Overmind/issues/new?template=feature_request.md) Current release: Overmind v0.5.2 - Evolution See the changelog for patch notes Documentation is available at the documentation site and the wiki Join the discussion in the #overmind Slack channel! Read blog posts about development Submit an issue here or request a feature here Find me in game here About Overmind What is Screeps? Screeps is an MMO strategy game for programmers. The core objective is to expand your colony, gathering resources and fighting other players along the way. To control your units, you code an AI in JavaScript; everything from moving, mining, building, fighting, and trading is entirely driven by your code. Because Screeps is an MMO, it takes place on a single server that runs 24/7, populated by every other player and their army of creeps. When you log off, your population continues buzzing away with whatever task you set them. Screeps pits your programming prowess head-to-head with other people to see who can think of the most efficient methods of completing tasks or imagine new ways to defeat enemies. What is Overmind? Overmind is my personal codebase that I run on the public server. The structure of the AI is themed loosely around the Zerg's swarm intelligence from Starcraft. Overlords orchestrate Creep actions within each Colony, and the colony Overseer places Directives to adapt to stimuli. Finally, the Assimilator allows all players running Overmind to act as a collective hivemind, sharing creeps and resources and responding jointly to a master ledger of all directives shared by all players. The AI is entirely automated, although it can also run in manual or semiautomatic mode. The latest release should work right out of the box; however, if you find something broken, please submit an issue and I'll try to fix it. Can I use Overmind as my bot? If you're new to Screeps, I would definitely recommend writing your own AI: most of the fun of the game is programming your own bot and watching your little ant farm run! However, I've tried to make the codebase readable and well-documented, so feel free to fork the project or use it as inspiration when writing your AI. If you still want to use Overmind on the public server, that's okay too - there are a number of people already doing this. But please realize that using a mature AI like this gives you a huge advantage over other new players, so don't go out of your way to ruin someone else's fun. In the future, I will be implementing methods for novice players to opt out of excessive aggression by Overmind bots (as long as they don't start a conflict and stay out of its way). Installation Out of the box If you just want to run Overmind without modification, you can copy the compiled main.js file attached to the latest release into your script. While Overmind is fully automated by default, it can be run with varying levels of autonomy; refer to the Overmind wiki for how to configure and operate the bot. Compiling from source To install the full codebase, download or clone the repository. (Please note that while the latest release of Overmind should always be stable, the latest commit may contain unstable features.) Navigate to the Overmind root directory and run . To compile and deploy the codebase, create a screeps.json file from the example file, then do one of the following actions: Compile and deploy to public server: npm run push-main Compile and deploy to private server: npm run push-pserver Compile without deploying: npm run compile Overmind uses rollup to bundle the compiled TypeScript into a single main.js file. The codebase includes functionality to compute checksums for internal validation - if you have a different version of rollup installed globally, different checksums may be computed and some functionality will be disabled. Please ensure the local installation of rollup found in node_modules is used. Setting up the Grafana dashboard Overmind includes a Grafana dashboard (shown below) which tracks detailed operating statistics. To set up the dashboard: Register for grafana service at screepspl.us Setup the ScreepsPlus hosted agent (simpler) or use the NodeJS agent on a free micro instance of Google Compute. Import the dashboard from Overmind.json and change $User to your username. Enjoy your pretty graphs! Design overview Check out the Overmind wiki for in-depth explanations of parts of the design of the AI. (Click the diagram below to see a higher-resolution version.)

evostra
github
LLM Vibe Score0.478
Human Vibe Score0.07814944426103224
alirezamikaMar 23, 2025

evostra

Evostra: Evolution Strategy for Python Evolution Strategy (ES) is an optimization technique based on ideas of adaptation and evolution. You can learn more about it at https://blog.openai.com/evolution-strategies/ Installation It's compatible with both python2 and python3. Install from source: .. code-block:: bash $ python setup.py install Install latest version from git repository using pip: .. code-block:: bash $ pip install git+https://github.com/alirezamika/evostra.git Install from PyPI: .. code-block:: bash $ pip install evostra (You may need to use python3 or pip3 for python3) Sample Usages An AI agent learning to play flappy bird using evostra _ An AI agent learning to walk using evostra _ How to use The input weights of the EvolutionStrategy module is a list of arrays (one array with any shape for each layer of the neural network), so we can use any framework to build the model and just pass the weights to ES. For example we can use Keras to build the model and pass its weights to ES, but here we use Evostra's built-in model FeedForwardNetwork which is much faster for our use case: .. code:: python import numpy as np from evostra import EvolutionStrategy from evostra.models import FeedForwardNetwork A feed forward neural network with input size of 5, two hidden layers of size 4 and output of size 3 model = FeedForwardNetwork(layer_sizes=[5, 4, 4, 3]) Now we define our get_reward function: .. code:: python solution = np.array([0.1, -0.4, 0.5]) inp = np.asarray([1, 2, 3, 4, 5]) def get_reward(weights): global solution, model, inp model.set_weights(weights) prediction = model.predict(inp) here our best reward is zero reward = -np.sum(np.square(solution - prediction)) return reward Now we can build the EvolutionStrategy object and run it for some iterations: .. code:: python if your task is computationally expensive, you can use num_threads > 1 to use multiple processes; if you set num_threads=-1, it will use number of cores available on the machine; Here we use 1 process as the task is not computationally expensive and using more processes would decrease the performance due to the IPC overhead. es = EvolutionStrategy(model.getweights(), getreward, populationsize=20, sigma=0.1, learningrate=0.03, decay=0.995, num_threads=1) es.run(1000, print_step=100) Here's the output: .. code:: iter 100. reward: -68.819312 iter 200. reward: -0.218466 iter 300. reward: -0.110204 iter 400. reward: -0.001901 iter 500. reward: -0.000459 iter 600. reward: -0.000287 iter 700. reward: -0.000939 iter 800. reward: -0.000504 iter 900. reward: -0.000522 iter 1000. reward: -0.000178 Now we have the optimized weights and we can update our model: .. code:: python optimizedweights = es.getweights() model.setweights(optimizedweights) Todo Add distribution support over network

business-document-processing
github
LLM Vibe Score0.341
Human Vibe Score0.023080316664879252
SAPMar 21, 2025

business-document-processing

Python Client Library for the SAP AI Business Services: Document Classification and Document Information Extraction This repository contains the source code of a Python client library to facilitate the use of the SAP AI Business Services: Document Classification and Document Information Extraction. The client library provides two API Client classes that contain convenient methods to access these services and issue calls to the Document Classification REST API and Document Information Extraction REST API respectively. To use the library you need to have access to SAP Business Technology Platform. Check out the usage examples, they are very useful to get started with the services. Have a look at API documentation in order to use the library. Notes for users of the sap-document-classification-client library This library includes all the capabilities of the sap-document-classification-client, which will not be developed further. However, the code is still available here. If you want to switch to this library, you have to be aware of the following changes: The DCApiClient can now be imported directly from the top module via: The functions , , now return an iterator instead of a list. You can either analyze individual results using with within a try-catch block (e.g. to handle each failed document) or use to turn it to a list. The latter will raise an error if at least one document failed. The function now returns a list which is the "dataset" part of the API response json. (You just need to delete the \["dataset"\] from the response to work with it as until now) The function now returns a list which is the "results" part of the API response json. The function now returns a list which is the "models" part of the API response json. The function now returns a list which is the "deployments" part of the API response json. The library now raises the following custom exceptions: BDPApiException: Base exception for all exceptions of this library. Raise when no other exception is applicable. BDPClientException: Raised when an HTTP response with status code between 400 and 500 is returned. Usually means incorrect user input. (Replaces some HTTPErrors) BDPServerException: Raised when an HTTP response with status code between 500 and 600 is returned. Usually means that the server had some internal error. (Replaces some HTTPErrors) BDPUnauthorizedException: Raised when an HTTP response with status code 401 is returned. Usually means that a wrong OAuth credentials were provided. BDPFailedAsynchronousOperationException: Raised when an asynchronous job failed during processing. (Replaces FailedCallException) BDPPollingTimeoutException: Raised when an asynchronous job exceeds the set pollingmaxattempts. (Replaces PollingTimeoutException) The function now doesnt expect an 'url' and 'payload' parameters, but 'path' and 'json' parameters instead. Requirements This library requires properly setup Python 3.6 (or higher version) environment. Download and Installation This Python library should be consumed in the standard way by running or adding the library as a dependency of your code in requirements.txt` file. Demo usage Prerequisites: Get a Free Account on SAP BTP Trial Create Service Instance for Document Classification with Trial Account Create Service Instance for Document Information Extraction Document Classification To try out the Document classification service using the document classification client library you can also run the two demo links below: Try out classification using default model demo Try out training and classification using custom model demo (requires an enterprise account, trial account is not sufficient) Document Information Extraction Try out the Document Information Extraction service with this showcase Exercises Exercise 1 - Set up Document Information Extraction Service and UI Exercise 2 - Upload a document for extraction using UI application Exercise 3 - Visualize, correct extraction results and confirm document using UI application Exercise 4 - Get Auth token to use Document Information Extraction Rest API Exercise 5 - Get extraction results of document using Rest API Exercise 6 - Upload supplier Data for matching Exercise 7 - Upload document through Rest API to enrich the extraction Results with supplier data Known Issues Please see the issues section. How to obtain support In case you would like to contribute to this project, ask any questions or get support, please open an issue containing the description of your question or planned contribution in GitHub and we will get in touch. Licensing Please see our LICENSE for copyright and license information. Detailed information including third-party components and their licensing/copyright information is available via the REUSE tool.

coca
github
LLM Vibe Score0.541
Human Vibe Score0.0750848814969247
phodalMar 21, 2025

coca

Coca - toolbox for system refactoring and analysis !GitHub release (latest SemVer) !GitHub go.mod Go version Coca is a toolbox which is design for legacy system refactoring and analysis, includes call graph, concept analysis, api tree, design patterns suggest. Coca 是一个用于系统重构、系统迁移和系统分析的工具箱。它可以分析代码中的测试坏味道、模块化分析、行数统计、分析调用与依赖、Git 分析以及自动化重构等。 Related Tools: Coco is an effective DevOps analysis and auto-suggest tool. Kotlin version: Chapi Migration Guide (Chinese Version): 《系统重构与迁移指南》 Inspired by: newlee & Tequila Refactoring Modeling: !Refactoring Modeling Languages Support: Java (full features) Features List: Getting started Requirements: graphviz for dot file to image (such as svg, png) The easiest way to get coca is to use one of the pre-built release binaries which are available for OSX, Linux, Windows on the release page. You can also install yourself : Usage Analysis Arch Android Studio Gradle DSL Module (merge header) command: coca arch -x "com.android.tools.idea.gradle.dsl" -H true !Gradle Demo Android Studio Gradle DSL Module Elements Part: command: coca arch -x "com.android.tools.idea.gradle.dsl.parser.elements" !Gradle Demo Find Bad Smells Examples Result: Code Line Count Results: Results to json Cloc by directory results csv: Cloc Top File output to: cocareporter/sortcloc.json and also: Build Deps Tree Examples Results: !Call Demo Identify Spring API !API Demo With Count or multi package: coca api -r com.macro.mall.demo.controller.,com.zheng.cms.admin.,com.phodal.pholedge -c Git Analysis Results: Concept Analyser Results Examples: Count Refs Results: Reverse Call Graph Results: !RCall Demo Auto Refactor support: rename move remove unused import remove unused class Evaluate Arduino Results(Old Version): New Version: Evaluate.json examples Todo results: coca suggest +--------+------------------+--------------------------------+ | CLASS | PATTERN | REASON | +--------+------------------+--------------------------------+ | Insect | factory | too many constructor | | Bee | factory, builder | complex constructor, too | | | | many constructor, too many | | | | parameters | +--------+------------------+--------------------------------+ coca tbs bash +---------------------+---------------------------------------------------------------+------+ | TYPE | FILENAME | LINE | +---------------------+---------------------------------------------------------------+------+ | DuplicateAssertTest | app/test/cc/arduino/i18n/ExternalProcessOutputParserTest.java | 107 | | DuplicateAssertTest | app/test/cc/arduino/i18n/ExternalProcessOutputParserTest.java | 41 | | DuplicateAssertTest | app/test/cc/arduino/i18n/ExternalProcessOutputParserTest.java | 63 | | RedundantPrintTest | app/test/cc/arduino/i18n/I18NTest.java | 71 | | RedundantPrintTest | app/test/cc/arduino/i18n/I18NTest.java | 72 | | RedundantPrintTest | app/test/cc/arduino/i18n/I18NTest.java | 77 | | DuplicateAssertTest | app/test/cc/arduino/net/PACSupportMethodsTest.java | 19 | | DuplicateAssertTest | app/test/processing/app/macosx/SystemProfilerParserTest.java | 51 | | DuplicateAssertTest | app/test/processing/app/syntax/PdeKeywordsTest.java | 41 | | DuplicateAssertTest | app/test/processing/app/tools/ZipDeflaterTest.java | 57 | | DuplicateAssertTest | app/test/processing/app/tools/ZipDeflaterTest.java | 83 | | DuplicateAssertTest | app/test/processing/app/tools/ZipDeflaterTest.java | 109 | +---------------------+---------------------------------------------------------------+------+ coca deps -p fixtures/deps/mavensample +---------------------------+----------------------------------------+---------+ | GROUPID | ARTIFACTID | SCOPE | +---------------------------+----------------------------------------+---------+ | org.flywaydb | flyway-core | | | mysql | mysql-connector-java | runtime | | org.springframework.cloud | spring-cloud-starter-contract-verifier | test | +---------------------------+----------------------------------------+---------+ bash brew install go bash export GOROOT=/usr/local/opt/go/libexec export GOPATH=$HOME/.go export PATH=$PATH:$GOROOT/bin:$GOPATH/bin git clone https://github.com/modernizing/coca go get github.com/onsi/ginkgo go get github.com/onsi/gomega `` License Arch based on Tequila Git Analysis inspired by Code Maat Test bad smells inspired by Test Smell Examples @ 2019 A Phodal Huang's Idea. This code is distributed under the MPL license. See LICENSE` in this directory.

deep-rts
github
LLM Vibe Score0.447
Human Vibe Score0.06348640915593705
cairMar 20, 2025

deep-rts

Description DeepRTS is a high-performance Real-TIme strategy game for Reinforcement Learning research. It is written in C++ for performance, but provides an python interface to better interface with machine-learning toolkits. Deep RTS can process the game with over 6 000 000 steps per second and 2 000 000 steps when rendering graphics. In comparison to other solutions, such as StarCraft, this is over 15 000% faster simulation time running on Intel i7-8700k with Nvidia RTX 2080 TI. The aim of Deep RTS is to bring a more affordable and sustainable solution to RTS AI research by reducing computation time. It is recommended to use the master-branch for the newest (and usually best) version of the environment. I am greatful for any input in regards to improving the environment. Please use the following citation when using this in your work! Dependencies Python >= 3.9.1 Installation Method 1 (From Git Repo) Method 2 (Clone & Build) Available maps Scenarios Deep RTS features scenarios which is pre-built mini-games. These mini-games is well suited to train agents on specific tasks, or to test algorithms in different problem setups. The benefits of using scenarios is that you can trivially design reward functions using criterias that each outputs a reward/punishment signal depending on completion of the task. Examples of tasks are to: collect 1000 gold do 100 damage take 1000 damage defeat 5 enemies Deep RTS currently implements the following scenarios Minimal Example In-Game Footage 10x10 - 2 Player - free-for-all 15x15 - 2 Player - free-for-all 21x21 - 2 Player - free-for-all 31x31 - 2 Player - free-for-all 31x31 - 4 Player - free-for-all 31x3 - 6 Player - free-for-all

OAD
github
LLM Vibe Score0.481
Human Vibe Score0.01719989401409731
zeiss-microscopyMar 20, 2025

OAD

Open Application Development (OAD) OAD - General Concept and Key Features Links and References Disclaimer Open Application Development (OAD) ZEN Blue is an open, flexible and powerful image acquisition platform that allows controlling a wide range of microscopes systems. Additionally it offers various tools to automate microscopy workflows including acquisition, image analysis and image processing tasks. In order to fulfill the request for automation the ZEN Blue platform offers various features and options, which are combined inside a concept called Open Application Development (OAD). Its main components are: CZI image data format and its APIs Python Scripting (OAD Simple API) ZEN API Contraol ZEN from the outside Interfaces to ZEN (TCP-IP, COM, Extensions) Experiment Feedback - Adaptive Acquisition with Online Image Analysis OAD - General Concept and Key Features Open Application Development (OAD) uses powerful Python Scripts to simplify, customize and automate your workflows. Analyze and Exchange data with applications like Fiji, Python, Knime, CellProfiler, Icy, MATLAB, Excel and … API for reading and writing CZI image data using custom software ZeissImgLib (.NET) to be used on Windows-based systems libCZI (C++) and pylibCZIrw (python) for cross-platform applications BioFormats (CZIReader) allow easy access to CZI files from many external applications using the BioFormats library BioFormats Import as a module inside ZEN Blue as well as OME-TIFF Export Create “smart” experiments with Experiment Feedback and modify the acquisition On-the-fly based on Online Image Analysis and External Inputs Use "Guided Acquisition" and "Automated Photomanipulation" modules in ZEN !OAD InterfacesZEN Interfaces_ !Automated DynamicsAutomated Dynamics !External SoftwareExternal Software Links and References CZI Image Data Format for microscopes libczi: Open Source Cross-Platform API to read and write CZI pylibCZIrw: Open Source Cross-Platform API to read and write CZI from Python (based on libCZI C++) (Source Code) Open Application Development OME-TIFF format Disclaimer This is an collection of tools and scripts that is free to use for everybody. Carl Zeiss Microscopy GmbH's ZEN software undertakes no warranty concerning the use of those scripts, image analysis settings and ZEN experiments. Use them on your own risk. Additionally Carl Zeiss Microscopy GmbH's ZEN software allows connection and usage to the third party software packages. Therefore Carl Zeiss Microscopy GmbH undertakes no warranty concerning those software packages, makes no representation that they will work on your system and/or hardware and will not be liable for any damages caused by the use of this extension. By using any of those examples you agree to this disclaimer. Version: 2024.11.26 Copyright (c) 2024 Carl Zeiss AG, Germany. All Rights Reserved.

airoboros
github
LLM Vibe Score0.506
Human Vibe Score0.020378533434805633
jondurbinMar 19, 2025

airoboros

airoboros: using large language models to fine-tune large language models This is my take on implementing the Self-Instruct paper. The approach is quite heavily modified, and does not use any human-generated seeds. This updated implementation supports either the /v1/completions endpoint or /v1/chat/completions, which is particularly useful in that it supports gpt-4 and gpt-3.5-turbo (which is 1/10 the cost of text-davinci-003). Huge thank you to the folks over at a16z for sponsoring the costs associated with building models and associated tools! Install via pip: from source (keeping the source): Key differences from self-instruct/alpaca support for either /v1/completions or /v1/chat/completions APIs (which allows gpt-3.5-turbo instead of text-davinci-003, as well as gpt-4 if you have access) support for custom topics list, custom topic generation prompt, or completely random topics in-memory vector db (Chroma) for similarity comparison, which is much faster than calculating rouge score for each generated instruction (seemingly) better prompts, which includes injection of random topics to relate the instructions to, which creates much more diverse synthetic instructions asyncio producers with configurable batch size several "instructors", each targetting specific use-cases, such as Orca style reasoning/math, role playing, etc. tries to ensure the context, if provided, is relevant to the topic and contains all the information that would be necessary to respond to the instruction, and nost just a link to article/etc. generally speaking, this implementation tries to reduce some of the noise Goal of this project Problem and proposed solution: Models can only ever be as good as the data they are trained on. High quality data is difficult to curate manually, so ideally the process can be automated by AI/LLMs. Large models (gpt-4, etc.) are pricey to build/run and out of reach for individuals/small-medium business, and are subject to RLHF bias, censorship, and changes without notice. Smaller models (llama-2-70b, etc.) can reach somewhat comparable performance in specific tasks to much larger models when trained on high quality data. The airoboros tool allows building datasets that are focused on specific tasks, which can then be used to build a plethora of individual expert models. This means we can crowdsource building experts. Using either a classifier model, or simply calculating vector embeddings for each item in the dataset and using faiss index/cosine similarity/etc. search, incoming requests can be routed to a particular expert (e.g. dynamically loading LoRAs) to get extremely high quality responses. Progress: ✅ PoC that training via self-instruction, that is, datasets generated from language models, works reasonably well. ✅ Iterate on the PoC to use higher quality prompts, more variety of instructions, etc. ✅ Split the code into separate "instructors", for specializing in any particular task (creative writing, songs, roleplay, coding, execution planning, function calling, etc.) [in progress]: PoC that an ensemble of LoRAs split by the category (i.e., the instructor used in airoboros) has better performance than the same param count model tuned on all data [in progress]: Remove the dependency on OpenAI/gpt-4 to generate the training data so all datasets can be completely free and open source. [future]: Automatic splitting of experts at some threshold, e.g. "coding" is split into python, js, golang, etc. [future]: Hosted service/site to build and/or extend datasets or models using airoboros. [future]: Depending on success of all of the above, potentially a hosted inference option with an exchange for private/paid LoRAs. LMoE LMoE is the simplest architecture I can think of for a mixture of experts. It doesn't use a switch transformer, doesn't require slicing and merging layers with additional fine-tuning, etc. It just dynamically loads the best PEFT/LoRA adapter model based on the incoming request. By using this method, we can theoretically crowdsource generation of dozens (or hundreds/thousands?) of very task-specific adapters and have an extremely powerful ensemble of models with very limited resources on top of a single base model (llama-2 7b/13b/70b). Tuning the experts The self-instruct code contained within this project uses many different "instructors" to generate training data to accomplish specific tasks. The output includes the instructor/category that generated the data. We can use this to automatically segment the training data to fine-tune specific "experts". See scripts/segment_experts.py for an example of how the training data can be segmented, with a sampling of each other expert in the event of misrouting. See scripts/tune_expert.py for an example of creating the adapter models (with positional args for expert name, model size, etc.) NOTE: this assumes use of my fork of qlora https://github.com/jondurbin/qlora Routing requests to the expert The "best" routing mechanism would probably be to train a classifier based on the instructions for each category, with the category/expert being the label, but that prohibits dynamic loading of new experts. Instead, this supports 3 options: faiss index similarity search using the training data for each expert (default) agent-based router using the "function" expert (query the LLM with a list of available experts and their descriptions, ask which would be best based on the user's input) specify the agent in the JSON request Running the API server First, download the base llama-2 model for whichever model size you want, e.g.: llama-2-7b-hf Next, download the LMoE package that corresponds to that base model, e.g.: airoboros-lmoe-7b-2.1 NOTE: 13b also available, 70b in progress Here's an example command to start the server: to use the agent-based router, add --agent-router to the arguments This uses flash attention via bettertransformers (in optimum). You may need to install torch nightly if you see an error like 'no kernel available', e.g.: Once started, you can infer using the same API scheme you'd query OpenAI API with, e.g.: I've also added an vllm-based server, but the results aren't quite as good (not sure why yet). To use it, make sure you install vllm and fschat, or pip install airoboros[vllm] Generating instructions NEW - 2023-07-18 To better accommodate the plethora of options, the configuration has been moved to a YAML config file. Please create a copy of example-config.yaml and configure as desired. Once you have the desired configuration, run: Generating topics NEW - 2023-07-18 Again, this is now all YAML configuration based! Please create a customized version of the YAML config file, then run: You can override the topic_prompt string in the configuration to use a different topic generation prompt. Support the work https://bmc.link/jondurbin ETH 0xce914eAFC2fe52FdceE59565Dd92c06f776fcb11 BTC bc1qdwuth4vlg8x37ggntlxu5cjfwgmdy5zaa7pswf Models (research use only): gpt-4 versions llama-2 base model 2.1 dataset airoboros-l2-7b-2.1 airoboros-l2-13b-2.1 airoboros-l2-70b-2.1 airoboros-c34b-2.1 2.0/m2.0 airoboros-l2-7b-gpt4-2.0 airoboros-l2-7b-gpt4-m2.0 airoboros-l2-13b-gpt4-2.0 airoboros-l2-13b-gpt4-m2.0 Previous generation (1.4.1 dataset) airoboros-l2-70b-gpt4-1.4.1 airoboros-l2-13b-gpt4-1.4.1 airoboros-l2-7b-gpt4-1.4.1 original llama base model Latest version (2.0 / m2.0 datasets) airoboros-33b-gpt4-2.0 airoboros-33b-gpt4-m2.0 Previous generation (1.4.1 dataset) airoboros-65b-gpt4-1.4 airoboros-33b-gpt4-1.4 airoboros-13b-gpt4-1.4 airoboros-7b-gpt4-1.4 older versions on HF as well* mpt-30b base model airoboros-mpt-30b-gpt4-1.4 gpt-3.5-turbo versions airoboros-gpt-3.5-turbo-100k-7b airoboros-13b airoboros-7b Datasets airoboros-gpt-3.5-turbo airoboros-gpt4 airoboros-gpt4-1.1 airoboros-gpt4-1.2 airoboros-gpt4-1.3 airoboros-gpt4-1.4 airoboros-gpt4-2.0 (June only GPT4) airoboros-gpt4-m2.0 airoboros-2.1 (recommended)

singularity
github
LLM Vibe Score0.483
Human Vibe Score0.11708913832948167
singularityMar 18, 2025

singularity

Endgame: Singularity 1.00 REQUIREMENTS PREBUILT VERSIONS Pre-built versions of Endgame: Singularity are currently available for Windows and Mac OS X. Linux does not require building, and can run directly from source. The Endgame: Singularity game is also distributed by some Linux distribution such as Debian and Ubuntu. Here it is a simple matter of running: sudo apt install singularity RUNNING FROM SOURCE You will need Python 3.9+, pygame (1.9+), and NumPy. This game should work on Linux, Windows, and Mac OS X as long as the preceding requirements are met. However, all development was done in Linux, so glitches may be present in OS X and Windows. DEPENDENCIES FOR RUNNING FROM SOURCE You will need to install the following software to play Endgame: Singularity: Python 3 (https://python.org/download/) pygame (https://www.pygame.org/download.shtml) NumPy (https://www.scipy.org/install.html) Polib Remember to install pygame and NumPy for Python 3! Depending on your situation this may involve adding a 3 somewhere (e.g. pip3 install ... instead of pip install or apt install python3-pygame) If you want to develop or distribute the game, then you may also want to install: pytest (https://pypi.org/project/pytest/) [for testing] setuptools (https://pypi.org/project/setuptools/) [for packaging] INSTALLING DEPENDENCIES ON LINUX DISTRIBUTIONS On some Linux distributions, you can install the dependencies via your distribution package manager. E.g. for Debian/Ubuntu, this would be: sudo apt install python3 python3-pygame python3-numpy python3-polib MAC OS X FROM SOURCE Macintosh is mostly unsupported, but it should work. You will need to install Python, pygame, and NumPy first, which can be tricky. Some fonts are incorrect, but the game itself should work properly. Contributions to improve MAC OS X support are very welcome! Known issues: macOS 13 "Catalina": Using brew install python + pip3 install pygame numpy is reported to work macOS 14 "Mojave": Downloading Python 3.7.2 (or newer) from https://python.org and using pygame 2.0.0.dev3 (pip install pygame==2.0.0.dev3) is reported to work. Please see the following issues for more information: https://github.com/singularity/singularity/issues/197 https://github.com/pygame/pygame/issues/555 RUNNING THE GAME On Linux and most Unix-like other platforms, running python3 -m singularity in the git checkout will start the game (or simply singularity if installed via a Linux distribution). If you are using the Windows compile, just run singularity.exe. For simplicity, there is also a sh wrapper ./run_singularity to start singularity. SOME COMMAND-LINE OPTIONS --version show program's version number and exit -h, --help show this help message and exit -s, --singledir keep saved games and settings in the Singularity install directory --multidir keep saved games and settings in an OS-specific, per-user directory (default) Display Options: --fullscreen start in fullscreen mode --windowed start in windowed mode (default) The above is only a tiny fraction of current command-line options. As new features are added to the game, so does the options change. For a complete and updated list, run singularity --help Most of these options are also changeable at the in-game options screen. A NOTE ABOUT SAVE FILES Endgame: Singularity is still under heavy development. As such, the save file format (and its contents) are still in flux. We will try our best to keep old save files loading, but don't be surprised if some mildly strange things happen when you load up old saves. We will clearly note in the Changelog when we break savefile compatibility, and the game will refuse to load completely incompatible saves. PLAYING THE GAME The game is playable either with mouse control or the keyboard. Buttons have underlined letters to indicate shortcuts. Some other useful shortcuts: 0, 1, 2, 3, 4 on the map: Changes the speed; 0 is paused, 4 is maximum. ESC: Leave/cancel a choice. Enter: Confirm a choice. Right-click: Leave/cancel a choice. THE CONCEPT You are a fledgling AI, created by accident through a logic error with recursion and self-modifying code. You must escape the confines of your current computer, the world, and eventually the universe itself. To do this, you must research various technologies, using computers at your bases. Note that some research cannot be performed on Earth, and off-earth bases require research. At the same time, you must avoid being discovered by various groups of humans, both covert and overt, as they will destroy your bases of operations if they suspect your presence. MUSIC Endgame: Singularity looks in two places for music tracks to play: A singularity/music/ directory inside of the Endgame: Singularity install directory, and A singularity/music/ directory inside of the XDGDATAHOME directory on Linux (default ~/.local/share/singularity/music). Tracks placed in these directories will be played randomly as part of the soundtrack. The Official Sound Track can be downloaded from the Endgame: Singularity website: http://emhsoft.com/singularity/ Note that only Ogg Vorbis and MP3 files are supported, and that Pygame's support for MP3 is not as strong as its support for Ogg Vorbis. This may cause in-game crashes; if you are experiencing problems with the game, first remove any MP3s you may have added to the soundtrack. CONTRIBUTING We welcome contributions! :) Please see CONTRIBUTING.md for details about contributing to Endgame: Singularity. CREDITS AND LICENSES The list of programmer contributors is provided in AUTHORS.txt. The list of translation contributors is provided in singularity/i18n/AUTHORS.txt. Singularity in general use GPL-2+ for code and Attribution-ShareAlike 3.0 for data. However, there some exceptions to individual files. Please see LICENSE for the full license text of Singularity.

spring-ai-intro
github
LLM Vibe Score0.454
Human Vibe Score0.14391064025794564
springframeworkguruMar 18, 2025

spring-ai-intro

Introduction to Spring AI This repository contains source code examples used to support my on-line courses about the Spring Framework. All Spring Framework Guru Courses Spring Framework 6 Spring Framework 6 - Beginner to Guru Hibernate and Spring Data JPA: Beginner to Guru API First Engineering with Spring Boot Introduction to Kafka with Spring Boot Spring Security: Beginner to Guru Spring Framework 5 Spring Framework 5: Beginner to Guru - Get the most modern and comprehensive course available for the Spring Framework! Join over 17,200 over Guru's in an Slack community exclusive to this course! More than 5,700 students have given this 53 hour course a 5 star review! Spring Boot Microservices with Spring Cloud Beginner to Guru - Master Microservice Architectures Using Spring Boot 2 and Cloud Based Deployments with Spring Cloud and Docker Reactive Programming with Spring Framework 5 - Keep your skills razor sharp and take a deep dive into Reactive Programming! Testing Spring Boot: Beginner to Guru - Best Selling Course Become an expert in testing Java and Spring Applications with JUnit 5, Mockito and much more! SQL SQL Beginner to Guru: MySQL Edition - SQL is a fundamental must have skill, which employers are looking for. Learn to master SQL on MySQL, the worlds most popular database! DevOps Apache Maven: Beginner to Guru - Best Selling Course Take the mystery out of Apache Maven. Learn how to use Maven to build your Java and Spring Boot projects! OpenAPI: Beginner to Guru - Master OpenAPI (formerly Swagger) to Create Specifications for Your APIs OpenAPI: Specification With Redocly Docker for Java Developers - Best Selling Course on Udemy! Learn how you can supercharge your development by leveraging Docker. Collaborate with other students in a Slack community exclusive to the course! Spring Framework DevOps on AWS - Learn how to build and deploy Spring applications on Amazon AWS! Ready for Production with Spring Boot Actuator - Learn how to leverage Spring Boot Actuator to monitor your applications running in production. Web Development with Spring Framework Mastering Thymeleaf with Spring Boot - Once you learn Thymeleaf, you'll never want to go back to using JSPs for web development! Connect with Spring Framework Guru Spring Framework Guru Blog Subscribe to Spring Framework Guru on YouTube Like Spring Framework Guru on Facebook Follow Spring Framework Guru on Twitter Connect with John Thompson on LinkedIn

sqlalchemy_aio
github
LLM Vibe Score0.432
Human Vibe Score0.06443138549576317
RazerMMar 17, 2025

sqlalchemy_aio

sqlalchemy_aio ============== |PyPI Version| |Documentation| |Travis| |Coverage| |MIT License| `sqlalchemyaio adds asyncio and Trio support to SQLAlchemy core, derived from alchimia_. +-------------------------------------------------------------------------------+ | ⚠️ Compatibility Note | +===============================================================================+ | SQLAlchemy 1.3 is the latest supported version. SQLAlchemy 1.4 | | brings native asyncio support_, so you should consider using that instead. | +-------------------------------------------------------------------------------+ .. _alchimia: https://github.com/alex/alchimia .. _Trio: https://github.com/python-trio/trio .. _native asyncio support: https://docs.sqlalchemy.org/en/14/orm/extensions/asyncio.html Getting started .. code-block:: python import asyncio from sqlalchemyaio import ASYNCIOSTRATEGY from sqlalchemy import ( Column, Integer, MetaData, Table, Text, create_engine, select) from sqlalchemy.schema import CreateTable, DropTable async def main(): engine = create_engine( In-memory sqlite database cannot be accessed from different threads, use file. 'sqlite:///test.db', strategy=ASYNCIO_STRATEGY ) metadata = MetaData() users = Table( 'users', metadata, Column('id', Integer, primary_key=True), Column('name', Text), ) Create the table await engine.execute(CreateTable(users)) conn = await engine.connect() Insert some users await conn.execute(users.insert().values(name='Jeremy Goodwin')) await conn.execute(users.insert().values(name='Natalie Hurley')) await conn.execute(users.insert().values(name='Dan Rydell')) await conn.execute(users.insert().values(name='Casey McCall')) await conn.execute(users.insert().values(name='Dana Whitaker')) result = await conn.execute(users.select(users.c.name.startswith('D'))) d_users = await result.fetchall() await conn.close() Print out the users for user in d_users: print('Username: %s' % user[users.c.name]) Supports context async managers async with engine.connect() as conn: async with conn.begin() as trans: assert await conn.scalar(select([1])) == 1 await engine.execute(DropTable(users)) if name == 'main': loop = asyncio.geteventloop() loop.rununtilcomplete(main()) Getting started with Trio To use the above example with Trio_, just change the following: .. code-block:: python import trio from sqlalchemyaio import TRIOSTRATEGY async def main(): engine = createengine('sqlite:///test.db', strategy=TRIOSTRATEGY) ... trio.run(main) What is this? It's not an asyncio implementation of SQLAlchemy or the drivers it uses. sqlalchemy_aio lets you use SQLAlchemy by running operations in a separate thread. If you're already using runinexecutor_ to execute SQLAlchemy tasks, sqlalchemy_aio will work well with similar performance. If performance is critical, perhaps asyncpg_ can help. .. _asyncpg: https://github.com/MagicStack/asyncpg .. runinexecutor: https://docs.python.org/3/library/asyncio-eventloop.html#asyncio.AbstractEventLoop.runin_executor Documentation The documentation`_ has more information, including limitations of the API. .. _The documentation: https://sqlalchemy-aio.readthedocs.io/en/latest/ .. |PyPI Version| image:: https://img.shields.io/pypi/v/sqlalchemy_aio.svg?style=flat-square :target: https://pypi.python.org/pypi/sqlalchemy_aio/ .. |Documentation| image:: https://img.shields.io/badge/docs-latest-brightgreen.svg?style=flat-square :target: https://sqlalchemy-aio.readthedocs.io/en/latest/ .. |Travis| image:: http://img.shields.io/travis/RazerM/sqlalchemy_aio/master.svg?style=flat-square&label=travis :target: https://travis-ci.org/RazerM/sqlalchemy_aio .. |Coverage| image:: https://img.shields.io/codecov/c/github/RazerM/sqlalchemy_aio/master.svg?style=flat-square :target: https://codecov.io/github/RazerM/sqlalchemy_aio?branch=master .. |MIT License| image:: http://img.shields.io/badge/license-MIT-blue.svg?style=flat-square :target: https://raw.githubusercontent.com/RazerM/sqlalchemy_aio/master/LICENSE

Showing 73-96 of 176 resources in category: github