# AI CEO Project Meeting Notes Date: Tuesday, January 28, 2025 ## Core Topics 1. Deep funding and dependency graphs analysis 2. Private vs. public information management for repositories and dependencies 3. MVP development strategy ## Contribution Evaluation Framework ### Types of Contributions #### Developer Contributions 1. Code metrics - Lines of code (with consideration for additions/deletions) - Pull requests (submitted, reviewed, merged) - Issue management (opened, closed) #### Social Contributions 1. Community support - Peer assistance - Question answering - Knowledge sharing 2. Content Creation - Educational materials - Documentation - Tutorial videos 3. Tipping System - Source Crowd implementation - Peer-to-peer value recognition - Tip-based contribution validation ### Tools & Platforms 1. Code Repository - GitHub integration - Metrics tracking - Contribution analysis 2. Social Platforms - Discord - Telegram - Communication tracking ## Evaluation Models ### Model Considerations 1. Rule-based Scoring (example) - Quantitative metrics (e.g., lines of code = 0.5 points) - PR complexity scoring (simple = 5 points, complex = 15 points) - Multiple model approach for comparison 2. Voting System - Model preference voting - Vote weighting mechanisms - Community consensus building ### Implementation Details 1. Automated Scripts - Database integration - Metrics collection - Real-time analysis 2. Review System - Peer review process - Contribution rating - Social component evaluation ## Challenges & Solutions ### Key Issues 1. Gaming Prevention - Cultural alignment - Anti-spam measures - Quality control 2. Incentive Alignment - Retroactive tipping index - Base plus tip model - Stake-based participation ### Risk Mitigation 1. Value Protection - Time-bound decay functions - Public/private information management - Smart contract implementation ## Framework Development ### Core Principles 1. Input Standardization - Agreed metrics - Data collection methods - Platform integration 2. Output Consistency - Token distribution - Value attribution - Performance measurement 3. Model Flexibility - Multiple model support - Customizable weights - Community-driven evolution ## Next Steps 1. LLM Integration - Unstructured to structured data conversion - Comment analysis - Feedback processing 2. Evaluation Schedule - Rop flight on Friday morning - Evaluation session scheduled for Friday ## Additional Considerations - Contextual information processing - Key decision maker identification - Weight distribution mechanisms - Compatibility assessment - Model testing and validation # *Granola based notes* ### Deep Funding and Contribution Evaluation - Discussed deep funding / dependency graphs and how private/public information on repos and dependencies work - Explored MVP (Minimum Viable Product) for evaluating contributions to emerging projects #### Types of Contributions: - Developer Contributions: - Debated weighting factors: number of lines of code, PRs - thinking: No ideal qualification of dev contribution exists - Proposed creating different models for community agreement - Doesn’t require training different models - Challenge: How to evaluate conversations on PRs - Rule-based Contributions: - Example: 1k lines of code added, X lines deleted, 4 PRs created, 4 PRs reviewed - Suggestion: Use a model to determine weights for different factors #### Tools and Data Sources: - GitHub: Code repository - Discord/Telegram: Communication platforms - Emphasized need for both code storage and discussion platforms #### Metrics to Consider: - Repository Metrics: - Lines of code (added/deleted) - PRs (created, reviewed, commented, merged) - Issues (opened, closed) - Social Metrics: - Quantifying contributions in discussions - Person-to-person help - General question answering - Content creation (e.g., educational videos) - Tipping system to indicate valuable contributions - Negative Contributions: - Spam - Disruptive behavior in social platforms ### Disagreements and Voting - Acknowledged potential disagreements on point systems - Proposed using multiple models (e.g., Model 1, Model 2) with different weightings - Suggested voting system: e.g., 4 votes for Model 1, 8 votes for Model 2 - Debated whether votes should be weighted ### Technical Implementation - Scripts to listen and pull data into a database - Reviewers/judges to rate PRs and models - First pass: Models provide list of contributions - Second pass: Random people with context rate contributions - Score models based on alignment with human ratings ### Challenges and Considerations - Handling private data in evaluation process - Retroactive vs. forward-looking evaluation - Balancing incentives and preventing gaming the system - Emergent project management and roadmap development - Competing priorities in evaluation ### Proposed Mechanisms - Index of liquid karitsu with retroactive tipping - Base plus tip model with pre-commitment - Options staking with rage quit for cash - Time-bound decay function for private/public information ### AI Model Considerations - Discussion on model behavior and preventing manipulation - Suggestion to model after improv techniques for more natural responses - Importance of breaking out of repetitive patterns ### Framework Agreement - Agree on inputs and outputs, disagree on actual model implementation - Use of GitHub and communication tools as data sources - Unified framework with flexibility for individual models ### Next Steps - Load and run the system with API keys - Implement multi-social network support - Develop method for key decision makers to provide private feedback - Test compatibility of envisioned system with various factors - Explore use of LLMs for unstructured to structured data conversion - Consider implementation of autonomous investor concept - Main track TBD, with evaluation on Friday --- ### Potential Gaming Behaviors 1. Message Spam - Rapid short messages to inflate message count - Breaking up single thoughts into multiple messages - Posting low-effort content like emoji reactions - Copy-pasting content across channels 2. Points/Reactions Manipulation - Coordinating with other users for mutual reaction farming - Creating alt accounts to give reactions - Strategically posting in channels with higher normalization factors - Excessive self-reporting of helping instances 3. User Activity Gaming - Creating artificial questions to "help" with - Inflating "questions asked" metric through minor variations - Cross-posting same content to maximize exposure - Time-based gaming (posting during low-activity periods to stand out) ### Suggested Countermeasures 1. Activity Pattern Detection - Implement rate limiting for messages - Track message length distribution and flag abnormal patterns - Monitor message uniqueness/similarity across channels - Track timing patterns between messages 2. User Behavior Analysis - Track reaction reciprocity patterns between users - Monitor help/question ratios for statistical anomalies - Implement cool-downs between point-earning activities - Add weights for message quality (length, complexity, uniqueness) 3. Structural Changes - Randomize or hide normalization factors - Add manual review for points above certain thresholds - Implement diminishing returns for rapid-fire activities - Create "quality scores" based on community engagement 4. Additional Metrics - Track edit/deletion patterns - Monitor user interaction networks for suspicious clusters - Implement sentiment analysis for message quality - Add time-based decay factors for points -- {%hackmd @xr/ai-comms %}