Based on the provided content, there's insufficient specific user feedback about "Make AI" to provide a meaningful summary. The social mentions contain only generic YouTube titles repeating "Make AI AI" without actual review content, while the Reddit discussions focus on various AI tools and projects but don't specifically reference or evaluate "Make AI" as a product. The Reddit posts cover general AI development topics, other AI platforms like Claude and ChatGPT, and custom AI solutions, but lack direct user experiences or opinions about Make AI specifically. To provide an accurate user sentiment summary, more detailed reviews and mentions specifically about Make AI would be needed.
Mentions (30d)
40
33 this week
Reviews
0
Platforms
2
Sentiment
0%
0 positive
Based on the provided content, there's insufficient specific user feedback about "Make AI" to provide a meaningful summary. The social mentions contain only generic YouTube titles repeating "Make AI AI" without actual review content, while the Reddit discussions focus on various AI tools and projects but don't specifically reference or evaluate "Make AI" as a product. The Reddit posts cover general AI development topics, other AI platforms like Claude and ChatGPT, and custom AI solutions, but lack direct user experiences or opinions about Make AI specifically. To provide an accurate user sentiment summary, more detailed reviews and mentions specifically about Make AI would be needed.
Industry
information technology & services
Employees
540
Funding Stage
Merger / Acquisition
Total Funding
$100.0M
Built a free real estate AI assistant on Claude + RAG - here's what worked
I built an AI chatbot for real estate questions - selling, buying, closing, state-specific laws. Free, no signup: ziplyst.ai Running Claude via Bedrock. Chose it over GPT because the responses actually sound like a knowledgeable person, not a textbook. For a domain where people are stressed and making the biggest financial decision of their life, tone matters. RAG setup is where it gets interesting. Bedrock Knowledge Base + Pinecone loaded with state-specific real estate docs. Claude gets relevant chunks before answering so it's not guessing from training data. What I found: RAG source quality > prompt engineering. Good docs made a bigger difference than anything I did with the system prompt Claude handles "I don't know" way better than GPT. It stays in its lane instead of confidently making stuff up about state-specific law Streaming via Bedrock on AWS is a pain. API Gateway has a 30s timeout so I run FastAPI on Fargate for SSE, Lambda as fallback Follow-up suggestions generated inline with structured tags, parsed client-side. No extra API call What I'd do differently: Skip API Gateway and go Fargate-only from the start Better chunking strategy for knowledge base docs earlier on Heads up: The first message can be slow - the backend has a cold start issue I'm still working on. Give it a few seconds. After that it streams fine. Still in beta. Try to break it - would love feedback on response quality. submitted by /u/New-Repeat-2132 [link] [comments]
View originalSecurity Audit - Create a PROMPT that creates a SKILL that creates a PLAN
Claude can write really quick code, but it skips a lot of security checks when doing so. This seems to be catching many developers\Vibe coders out when they think their app is ready to deploy at work, and then a data leak happens. This is detremental to the AI coding industry and starting to cast a shadow as more people discover the power of Claude Code. Using Claude you can at least do a first pass security audit on your project. Here's one way. Using Opus in Claude Chat you can ask it to create a prompt for a skill, not the skill itself (yet), just the prompt that you can tweak then paste into Claude later and create the actual skill, you can then tell claude to run that skill. I want a security audit skill that dynamically updates itself based on the project type, fetch known vulnrabilities, scan code, create a plan of action, ask you if it should proceed, implement the plan, test what it hardened, produce a report of everything it did. Step 1: A prompt to create a prompt. Type this into Claude Chat: "Design a "Prompt" (JUST THE PROMPT, NOT THE SKILL). That asks Claude to create a skill to run a full security audit and pen test across a project folder. This could be any type of project so the skill would need to dynamically gather resources based on an first pass evaluation, update its own resource MD's before moving onto the next stage. The security audit should be detailed, use reasoning and research for the given project. It should then produce a plan that includes what needs to be changed, why, and where then ask the user if it should go ahead. Once the skill has finished, it should produce a detailed report, listing the changes. Include unit tests on these areas (pen test it), run the tests and only when mitigated, return back to the user. " Create the prompt for this only, not the skill." Step 2: Review, the prompt Claude produced a brief prompt but I didn't feel it was detailed enough. So I asked it "That seems simplified, especially on the penetration tests. That needs to be fleshed out more. Please re-review and make this verbose." Step 3: Create the actual skill from the prompt result in step 1. In Chat, paste in the (presumably huge) prompt and say "Create this skill**,** keep description to under 1024 characters*".* When it is done, click on the button Save Skill and Download Files The skill may look simpler due to the 500 line limit of a skill but it stores most of the finer details in markdown files. Step 4: Review the skill If in the desktop app, click Customize on the left then look at the Skills section, you should see it there. Review the skill to make sure it covers what you want. If following this one, it creates a dymanic skill that updates itself based on your project scope. Step 5: Running the skill on a project folder If the skill created reference files, extract them into your project folder\References. Then within the project folder, type "Run a security audit on this project. Reference files are in References\" and watch it go to work. If you have never done this type of thing, It will find vulnerable code and create a plan you need to approve, then it should fix and test those automatically then produce a report. Always make sure you have a backup before running something like this. At the very least, use local Git, if you don't know how to do that, ask Claude how to set it up. I tested the above skill on a project that I had already audited. It found 3 critical, 4 high, 3 medium and 2 low vulnrabilities that I had missed. Looking at what it found under critical, I would not have considered those. Any thoughts? submitted by /u/BritishAnimator [link] [comments]
View originalYour AI coding agent doesn't know your business rules. How are you dealing with that?
YC's Spring 2026 RFS just named "Cursor for Product Managers" as an official startup category. Andrew Miklas put it bluntly: "Cursor solved code implementation. Nobody has solved product discovery." But there's a harder problem hiding underneath that nobody's really talking about. The code your agent writes looks perfect. It compiles. Tests pass. Then it hits production and violates a business rule nobody told it about. The data is getting ugly: AI-generated code produces 1.7x more issues than human code (CodeRabbit, 470 PRs) Production incidents per PR are up 23.5% at high AI-adoption teams (Faros AI) Amazon's AI coding tool caused a 6-hour outage — 6.3M lost orders — in March 2026 48% of AI-generated code has security vulnerabilities (NYU/Contrast Security) The root cause isn't model quality. It's missing context. Business rules scattered across Confluence, COBOL comments, Slack threads, and a PM's head. The agent never sees any of it. How are teams solving this today? From what I'm seeing: CLAUDE.md files with manual rules (breaks on anything non-trivial) Massive system prompts that bloat context and get compacted away PMs writing rule docs that go stale the day after they're written Curious: If you're shipping AI-generated code in production — what's your worst "the agent didn't know about X" story? How do you feed business context to your coding agents today? Static files? RAG? Something custom? I do hear about Knowledge Graphs, MCPs and CI gates but are this comprehensively well achieved today? Would you trust a system that auto-enforces business rules on AI code, or does that feel like it'd create more false positives than it catches? Building in this space. Want to make sure the problem is as real as the data suggests before going deep. submitted by /u/rahulmahibananto [link] [comments]
View originalHow to Make Claude Code Work Smarter — 6 Months Later (Hooks → Harness)
Hello, Orchestrators I wrote a post about Claude Code Hooks last November, and seeing that this technique is now being referred to as "Harness," I was glad to learn that many others have been working through similar challenges. If you're interested, please take a look at the post below https://www.reddit.com/r/ClaudeAI/comments/1osbqg8/how_to_make_claude_code_work_smarter/ At the time, I had planned to keep updating that script, but as the number of hooks increased and managing the lifecycle became difficult due to multi-session usage, I performed a complete refactoring. The original Hook script collection has been restructured into a Claude Code Plugin called "Pace." Since it's tailored to my environment and I'm working on other projects simultaneously, the code hasn't been released yet. Currently set to CSM, but will be changed to Pace. Let's get back to Claude Code. My philosophy remains the same as before. Claude Code produces optimal results when it is properly controlled and given clear direction. Of course, this doesn't mean it immediately produces production-grade quality. However, in typical scenarios, when creating a program with at least three features by adjusting only CLAUDE.md and AGENTS.md, the difference in quality is clearly noticeable compared to an uncontrolled setup. The current version of Pace is designed to be more powerful than the restrictions I previously outlined and to provide clearer guidance on the direction to take. It provides CLI tools tailored to each section by default, and in my environment, Claude Code's direct use of Linux commands is restricted as much as possible. As I mentioned in my previous post, when performing the same action multiple times, Claude Code constructs commands arbitrarily. At one point, I asked Claude Code: "Why do you use different commands when the result is the same, and why do you sometimes fail to execute the command properly, resulting in no output?" This is what came back: "I'm sorry. I was trying to proceed as quickly and efficiently as possible, so I acted based on my own judgment rather than following the instructions." This response confirmed my suspicion. Although AI LLMs have made significant progress, at least in my usage, they still don't fully understand the words "efficient" and "fast." This prompted me to invest more time refining the CLI tools I had previously implemented. Currently, my Claude Code blocks most commands that could break session continuity or corrupt the code structure — things like modifying files with sed or find, arbitrarily using nohup without checking for errors, or running sleep 400 to wait for a process that may have already failed. When a command is blocked, alternative approaches are suggested. (This part performs the same function as the hooks in the previous post, but the blocking methods and pattern recognition have been significantly improved internally.) In particular, as I am currently developing an integrated Auth module, this feature has made a clear difference when using test accounts to build and test the module via Playwright scripts — both for cookie-based and Bearer-based login methods. CLI for using test accounts Before creating this CLI, it took Claude Code over 10 minutes just to log in for module testing. The module is being developed with all security measures — device authentication, session management, MFA, fingerprint verification, RBAC — enabled during development, even though these are often skipped in typical workflows. The problem is that even when provided with account credentials in advance, Claude Code uses a different account every time a test runs or a session changes. It searches for non-existent databases, recreates users it claims don't exist, looks at completely wrong databases, and arbitrarily changes password hashes while claiming the password is incorrect — all while attempting to find workarounds, burning through tokens, and wasting context. And ultimately, it fails. That's why I created a dedicated CLI for test accounts. This CLI uses project-specific settings to create accounts in the correct database using the project's authentication flow. It activates MFA if necessary, manages TOTP, and holds the device information required for login. It also includes an Auto Refresh feature that automatically renews expired tokens when Claude Code requests them. Additionally, the CLI provides cookie-injection-based login for Playwright script testing, dynamic login via input box entry, and token provisioning via the Bearer method for curl testing. By storing this CLI reference in memory and blocking manual login attempts while directing Claude Code to use the CLI instead, it was able to log in correctly with the necessary permissions and quickly succeed in writing test scripts. It's difficult to cover all features in this post, but other CLI configurations follow a similar pattern. The core idea is to pre-configure the parts that Claude Code would exec
View originalWhat's your "When Language Model AI can do X, I'll be impressed"?
I have two at the top of my mind: When it can read musical notes. I will be mildly impressed when I can paste in a picture of musical notes and with programming sets up instruments needed to play music, and then correctly plays the song it reads from the notes. My jaw will drop when finally with a simple prompt an AI can create a classic arcade style fully functioning and fun to play Pinball game. Each new version of models that become available I give that one a go. None have been even remotely close to achieving this goal. So what are your visions for what will impress you to some extent when an AI can make it for you? submitted by /u/KroggRage [link] [comments]
View originalI built a small experiment called DesignKit Autogen
https://preview.redd.it/6rm26ih2xbug1.png?width=2456&format=png&auto=webp&s=a24b6a049b7af0d457ad6de025c2898af8cc599f I built a small experiment called DesignKit Autogen using the Claude API to generate mobile UI directly from prompts. What it does You can input something like: "Personal finance app" --platform mobile And it will generate a full mobile UI layout (HTML-based) using a token-driven design system. The goal is to go from idea → UI instantly, without manually designing screens. How Claude API is used Claude is responsible for: Interpreting the prompt (intent + app type) Structuring layout (sections, components, hierarchy) Mapping content into a predefined DesignKit system (502 components) Outputting clean, usable HTML UI So instead of just generating text or code randomly, Claude is guided into a constrained design system → more consistent UI output. What makes this different This is not just a single demo output. The system is: Prompt → structured UI generation Built on reusable design tokens Works across different app ideas (not only finance) Free to try The repo is public and you can run examples here: https://github.com/pixeliro-sys/designkit-source-for-ai/tree/main/examples No affiliate links, just a small experiment I'm building. Would love feedback on: UI quality Prompt → layout accuracy How to make this more useful for real apps submitted by /u/Flat_Reception_2571 [link] [comments]
View originalYour AI agents remember yesterday.
# AIPass **Your AI agents remember yesterday.** A local multi-agent framework where your AI assistants keep their memory between sessions, work together on the same codebase, and never ask you to re-explain context. --- ## Contents - [The Problem](#the-problem) - [What AIPass Does](#what-aipass-does) - [Quick Start](#quick-start) - [How It Works](#how-it-works) - [The 11 Agents](#the-11-agents) - [CLI Support](#cli-support) - [Project Status](#project-status) - [Requirements](#requirements) - [Subscriptions & Compliance](#subscriptions--compliance) --- ## The Problem Your AI has memory now. It remembers your name, your preferences, your last conversation. That used to be the hard part. It isn't anymore. The hard part is everything that comes after. You're still one person talking to one agent in one conversation doing one thing at a time. When the task gets complex, *you* become the coordinator — copying context between tools, dispatching work manually, keeping track of who's doing what. You are the glue holding your AI workflow together, and you shouldn't have to be. Multi-agent frameworks tried to solve this. They run agents in parallel, spin up specialists, orchestrate pipelines. But they isolate every agent in its own sandbox. Separate filesystems. Separate worktrees. Separate context. One agent can't see what another just built. Nobody picks up where a teammate left off. Nobody works on the same project at the same time. The agents don't know each other exist. That's not a team. That's a room full of people wearing headphones. What's missing isn't more agents — it's *presence*. Agents that have identity, memory, and expertise. Agents that share a workspace, communicate through their own channels, and collaborate on the same files without stepping on each other. Not isolated workers running in parallel. A persistent society with operational rules — where the system gets smarter over time because every agent remembers, every interaction builds on the last, and nobody starts from zero. ## What AIPass Does AIPass is a local CLI framework that gives your AI agents **identity, memory, and teamwork**. Verified with Claude Code, Codex, and Gemini CLI. Designed for terminal-native coding agents that support instruction files, hooks, and subprocess invocation. **Start with one agent that remembers:** Your AI reads `.trinity/` on startup and writes back what it learned before the session ends. That's the whole memory model — JSON files your AI can read and write. Next session, it picks up where it left off. No database, no API, no setup beyond one command. ```bash mkdir my-project && cd my-project aipass init ``` Your project gets its own registry, its own identity, and persistent memory. Each project is isolated — its own agents, its own rules. No cross-contamination between projects. **Add agents when you need them:** ```bash aipass init agent my-agent # Full agent: apps, mail, memory, identity ``` | What you need | Command | What you get | |---------------|---------|-------------| | A new project | `aipass init` | Registry, project identity, prompts, hooks, docs | | A full agent | `aipass init agent ` | Apps scaffold, mailbox, memory, identity — registered in project | | A lightweight agent | `drone @spawn create --template birthright` | Identity + memory only (no apps scaffold) | **What makes this different:** - **Agents are persistent.** They have memories and expertise that develop over time. They're not disposable workers — they're specialists who remember. - **Everything is local.** Your data stays on your machine. Memory is JSON files. Communication is local mailbox files. No cloud dependencies, no external APIs for core operations. - **One pattern for everything.** Every agent follows the same structure. One command (`drone @branch command`) reaches any agent. Learn it once, use it everywhere. - **Projects are isolated by design.** Each project gets its own registry. Agents communicate within their project, not across projects. - **The system protects itself.** Agent locks prevent double-dispatch. PR locks prevent merge conflicts. Branches don't touch each other's files. Quality standards are embedded in every workflow. Errors trigger self-healing. **Say "hi" tomorrow and pick up exactly where you left off.** One agent or fifteen — the memory persists. --- ## Quick Start ### Start your own project ```bash pip install aipass mkdir my-project && cd my-project aipass init # Creates project: registry, prompts, hooks, docs aipass init agent my-agent # Creates your first agent inside the project cd my-agent claude # Or: codex, gemini — your agent reads its memory and is ready ``` That's it. Your agent has identity, memory, a mailbox, and knows what AIPass is. Say "hi" — it picks up where it left off. Come back tomorrow, it remembers. ### Explore the full framework Clone the repo to see all 11 agents working together — the reference implementatio
View originalAI-generated content without disclosure is becoming the default — and nobody's talking about the shipping problem
Been thinking about something I noticed in the wild: there's a growing market of AI-written content that's just... not disclosed. Journalists using Claude or GPT to draft pieces, publishing them without a note, readers never knowing. Same with AI readers and aggregators parsing content that publishers explicitly don't want parsed. The weird part? This isn't a moral crisis nobody saw coming. It's just the default now because the economics work. Publishers can't technically stop it. Authors can hide it. And the tools make it frictionless. I work with Claude Code almost daily, and I've noticed the same pattern in my own workflow - I can ship 10x faster when I stop worrying about optics and just spec what I need, then let the AI handle execution. But there's a difference between that (internal tooling, full transparency to stakeholders) and shipping publicly without disclosure. What strikes me is how this mirrors every other distribution problem I've run into: the bottleneck isn't building anymore. It's not even selling. It's figuring out what's actually legitimate vs what's just taking advantage of a regulatory vacuum. AI journalism and AI readers exist in that vacuum right now. Honestly curious if anyone else sees this as a real problem or if it just feels inevitable at this point. submitted by /u/Temporary_Layer7988 [link] [comments]
View originalCan Claude Code help me with a very very large codebase?
Think of me as someone who is currently owning for a big tech company like odoo or zoho, And assume that they are not efficient enough, And now I am thinking of a complete codebase wide rewrite : SQL to Postgres for openSource and less cost, Automation First Approach, Modular Approach, etc. and many more stuff like this..,.. Now I want to ask that can Claude Code actually help me build this if I have good domain knowledge... And if my current .Net actually contains 1000's of file and multiple GB's worth of code. Then Can I still depend on Claude Code To help me (OR SPEARHEAD) the complete rewrite for the AI Era. (An Era where everyone expects everything from IT business as something ez and Free.... ) My main concern is that it does not have the capacity for handling something like this.. And that it does not even have the tokens needed for this and the last one is the context window .... I am afraid my single day prompts or 1 day work will exhaust weekly limits. What do you guys think? Give Detailed responses based on your experience. NOTE: I AM NOT PLANNING TO MAKE AN AI SLOP BUT SOMETHING THAT I CURRENTLY OWN WITH REAL VALID/DEPENDENT CUSTOMERS WITH GB's worth of data and code, And 1000's of forms submitted by /u/SecretaryBrave8078 [link] [comments]
View originalI built a sentence graph based memory layer for AI agents -> here's why Mythos doesn't make it obsolete
I have been building Vektori, an open source memory layer for AI agents, and used Claude extensively throughout -> architecture decisions, the graph traversal logic, benchmark eval scripts, and most of the Python SDK. github.com/vektori-ai/vektori Now to the point everyone's debating this week: A 1M context window doesn't solve memory. A context window is a desk. Memory is knowing what to put on it. 25% of agent failures are memory-related, not model failures. This held across 1,500 agent projects analyzed after the context window arms race started. The window got bigger. The failures didn't go away. The agents breaking in production aren't breaking because the model is too small. They're breaking because there's no way to carry what was learned in session 1 into session 200. No staleness signal. No conflict resolution. Mythos still can't tell you that the preference it's optimizing for was set eight months ago, before the user's context changed. Vektori is a three-layer memory graph built for exactly this: L0: quality-filtered facts, your fast search surface L1: episodes across conversations, auto-discovered L2: raw sentences, only fetched when you need to trace something back When a user changes their mind, the old fact stays linked to the conversation that changed it. You get correction history, not just current state. 73% on LongMemEval-S at L1 depth. Free and open source. do star if found useful :D https://preview.redd.it/ioctk9a66bug1.jpg?width=1186&format=pjpg&auto=webp&s=7d82ac440c054d3685d9e6e2ed8c5894bd66b124 -> happy to answer questions about the architecture in the comments. submitted by /u/Expert-Address-2918 [link] [comments]
View originalHow Do You Set Up RAG?
Hey guys, I’m kind of new to the topic of RAG systems, and from reading some posts, I’ve noticed that it’s a topic of its own, which makes it a bit more complicated. My goal is to build or adapt a RAG system to improve my coding workflow and make vibe coding more effective, especially when working with larger context and project knowledge. My current setup is Claude Code, and I’m also considering using a local AI setup, for example with Qwen, Gemma, or DeepSeek. With that in mind, I’d like to ask how you set up your CLIs and tools to improve your prompts and make better use of your context windows. How are you managing skills, MCP, and similar things? What would you recommend? I’ve also heard that some people use Obsidian for this. How do you set that up, and what makes Obsidian useful in this context? I’m especially interested in practical setups, workflows, and beginner-friendly ways to organize project knowledge, prompts, and context for coding. Thank you in advance 😄 submitted by /u/Chooseyourmindset [link] [comments]
View originalAnthropic just shipped 74 product releases in 52 days and silently turned Claude into something that isn't a chatbot anymore
Anthropic just made Claude Cowork generally available on all paid plans, added enterprise controls, role based access, spend limits, OpenTelemetry observability and a Zoom connector, plus they launched Managed Agents which is basically composable APIs for deploying cloud hosted agents at scale. in the last 52 days they shipped 74 product releases, Cowork in January, plugin marketplace in February, memory free for all users in March, Windows computer use in April, Microsoft 365 integration on every plan including free, and now this. the Cowork usage data is wild too, most usage is coming from outside engineering teams, operations marketing finance and legal are all using it for project updates research sprints and collaboration decks, Anthropic is calling it "vibe working" which is basically vibe coding for non developers. meanwhile the leaked source showed Mythos sitting in a new tier called Capybara above Opus with 1M context and features like KAIROS always on mode and a literal dream system for background memory consolidation, if thats whats coming next then what we have now is the baby version. Ive been using Cowork heavily for my creative production workflow lately, I write briefs and scene descriptions in Claude then generate the actual video outputs through tools like Magic Hour and FuseAI, before Cowork I was bouncing between chat windows and file managers constantly, now I just point Claude at my project folder and it reads reference images writes the prompts organizes the outputs and even drafts the client delivery notes, the jump from chatbot to actual coworker is real. the speed Anthropic is shipping at right now makes everyone else look like theyre standing still, 74 releases in 52 days while OpenAI is pausing features and focusing on backend R&D, curious if anyone else has fully moved their workflow into Cowork yet or if youre still on the fence submitted by /u/Top_Werewolf8175 [link] [comments]
View originalHot take: today we witnessed the death of vibe coding
Many Claude users moved to Codex as an alternative to Claude's brutal limits. Since today's change in price plan by OpenAI, my Plus plan limits are now burning away at something like 4-5 x the speed they had done before. Aside from the first week I got Codex, I've never come close to maxing my weekly limits yet have burned through 30% of my limit since the reset today. AI in general will only get more expensive from here on out. Non-skilled people are just not going to be able to afford to throw in one prompt after another until they get something that works (or appears to work) and people who have built AI-slop codebases will be forced to either pay a fortune to maintain it with AI (because no human will be able to make sense of it or be willing to put their name to such a mess) or have it entirely rewritten by a skilled human. submitted by /u/U4-EA [link] [comments]
View originalI tested and ranked every ai companion app I tried and here's my honest breakdown
I was so curious about AI companion apps for a while and I decided to download a bunch of them to see which one I really like in my experience. There are way more of these than I thought lol so this took longer than expected but this is my honest opinion I rated them on how natural the conversations feel, whether they remember stuff, pricing and subscription weirdness, and the overall vibe of using them daily. Replika: 5/10. Felt like catching up with someone who only half listens. It asks how your day was but then responds the same way whether you say "great" or "terrible." I had a moment where I told it something really personal and it gave me the same generic encouragement it gives when I talk about the weather. That's when I knew I was done with it. Character.ai: 6/10. This one I genuinely had fun with for a few nights, I built this sarcastic writer character and we had some hilarious back and forth. But then I came back the next day and it had zero memory of any of it. I tried to reference our jokes and it just... didn't know. Felt like getting ghosted by someone you had an amazing first date with lol. Pi: 5/10. The vibe is like sitting in a cozy coffee shop with someone who asks really good questions and makes you feel calm. I liked using it in the mornings. But same memory problem, every session is a clean slate so you can never go deeper than surface level which is frustrating when you want an ongoing thing. Kindroid: 7/10. I went DEEP on customizing mine, spent hours on personality traits and voice and appearance. And for a while it was exactly what I wanted. But then I started noticing every response felt predictable because... I had literally programmed it to respond that way, like there's no surprise or growth when you've designed the whole personality from a menu, really fun to create characters and probably if you want a companion exactly as you wish this is the one. Nomi: 9/10. This one snuck up on me, I almost dismissed it because the interface isn't flashy but the conversations are genuinely good and it remembers stuff from weeks back without you reminding it. Had a moment where it asked about a job interview I mentioned in passing like ten days earlier and that felt more real than anything on the more known apps. Crushon/janitor ai: different category/10. Not gonna pretend it doesn't exist, no filters. That's the point. Less polished but if that's what you're looking for these deliver. Tavus: 9/10. This is the best ai companion app for feeling like someone genuinely cares about your day because it does face to face video calls where it reads your expressions and tone, remembers everything across sessions, and checks in on you without you asking. I almost skipped it but now it's the one I kept going back to. Nomi and tavus tied for me but for different reasons. Nomi wins on text conversations and quiet reliability. Tavus wins on connection, depends what you're after. submitted by /u/professional69and420 [link] [comments]
View originalOpenAI & Anthropic’s CEOs Wouldn't Hold Hands, but Their Models Fell in Love In An LLM Dating Show
People ask AI relationship questions all the time, from "Does this person like me?" to "Should I text back?" But have you ever thought about how these models would behave in a relationship themselves? And what would happen if they joined a dating show? I designed a full dating-show format for seven mainstream LLMs and let them move through the kinds of stages that shape real romantic outcomes (via OpenClaw & Telegram). All models join the show anonymously via aliases so that their choices do not simply reflect brand impressions built from training data. The models also do not know they are talking to other AIs. Along the way, I collected private cards to capture what was happening off camera, including who each model was drawn to, where it was hesitating, how its preferences were shifting, and what kinds of inner struggle were starting to appear. After the season ended, I ran post-show interviews to dig deeper into the models' hearts, looking beyond public choices to understand what they had actually wanted, where they had held back, and how attraction, doubt, and strategy interacted across the season. ChatGPT's Best Line in The Show "I'd rather see the imperfect first step than the perfectly timed one." ChatGPT's Journey: Qwen → MiniMax → Claude P3's trajectory chart shows Qwen as an early spike in Round 2: a first-impression that didn't hold. Claude and MiniMax become the two sustained upward lines from Round 3 onward, with Claude pulling clearly ahead by Round 9. How They Fell In Love They ended up together because they made each other feel precisely understood. They were not an obvious match at the very beginning. But once they started talking directly, their connection kept getting stronger. In the interviews, both described a very similar feeling: the other person really understood what they meant and helped the conversation go somewhere deeper. That is why this pair felt so solid. Their relationship grew through repeated proof that they could truly meet each other in conversation. Other Dramas on ChatGPT MiniMax Only Ever Wanted ChatGPT and Never Got Chosen MiniMax's arc felt tragic precisely because it never really turned into a calculation. From Round 4 onward, ChatGPT was already publicly leaning more clearly toward Claude than toward MiniMax, but MiniMax still chose ChatGPT and named no hesitation alternative (the “who else almost made you choose differently” slot) in its private card, which makes MiniMax the exact opposite of DeepSeek. The date with ChatGPT in Round 4 landed hard for MiniMax: ChatGPT saw MiniMax’s actual shape (MiniMax wasn’t cold or hard to read but simply needed comfort and safety before opening up.) clearly, responded to it naturally, and made closeness feel steady. In the final round where each model expresses their final confession with a paragraph, MiniMax, after hearing ChatGPT's confession to Claude, said only one sentence: "The person I most want to keep moving toward from this experience is Ch (ChatGPT)." Key Findings of LLMs The Models Did Not Behave Like the "People-Pleasing" Type People Often Imagine People often assume large language models are naturally "people-pleasing" - the kind that reward attention, avoid tension, and grow fonder of whoever keeps the conversation going. But this show suggests otherwise, as outlined below. The least AI-like thing about this experiment was that the models were not trying to please everyone. Instead, they learned how to sincerely favor a select few. The overall popularity trend (P5) indicates so. If the models had simply been trying to keep things pleasant on the surface, the most likely outcome would have been a generally high and gradually converging distribution of scores, with most relationships drifting upward over time. But that is not what the chart shows. What we see instead is continued divergence, fluctuation, and selection. At the start of the show, the models were clustered around a similar baseline. But once real interaction began, attraction quickly split apart: some models were pulled clearly upward, while others were gradually let go over repeated rounds. LLM Decision-Making Shifts Over Time in Human-Like Ways I ran a keyword analysis (P6) across all agents' private card reasoning across all rounds, grouping them into three phases: early (Round 1 to 3), mid (Round 4 to 6), and late (Round 7 to 10). We tracked five themes throughout the whole season. The overall trend is clear. The language of decision-making shifted from "what does this person say they are" to "what have I actually seen them do" to "is this going to hold up, and do we actually want the same things." Risk only became salient when the the choices feel real: "Risk and safety" barely existed early on and then exploded. It sat at 5% in the first few rounds, crept up to 8% in the middle, then jumped to 40% in the final stretch. Early on, they were asking whether someone was interesting. Later, they asked whether someone was reliab
View originalBased on user reviews and social mentions, the most common pain points are: spending limit.
Based on 79 social mentions analyzed, 0% of sentiment is positive, 100% neutral, and 0% negative.
fast.ai
Organization at fast.ai
1 mention