/robowaifu/ - DIY Robot Wives

Advancing robotics to a point where anime catgrill meidos in tiny miniskirts are a reality.

Site was down because of hosting-related issues. Figuring out why it happened now.

Build Back Better

Sorry for the delays in the BBB plan. An update will be issued in the thread soon in late August. -r

Max message length: 6144

Drag files to upload or
click here to select them

Maximum 5 files / Maximum size: 20.00 MB

More

(used to delete files and postings)


“I am not judged by the number of times I fail, but by the number of times I succeed: and the number of times I succeed is in direct proportion to the number of times I fail and keep trying.” -t. Tom Hopkins


Visual Waifus Robowaifu Technician 09/15/2019 (Sun) 06:40:42 No.240
Thoughts on waifus which remain 2D but have their own dedicated hardware. This is more on the artistry side though ai is still involved. An example of an actual waifu product being the Gatebox.
gatebox.ai/sp/

My favorite example is Ritsu, she's a cute ai from assassination classroom who's body is a giant screen on wheels.
>>21406 I did watch the video and looked at the links, but it doesn't explain what to do in a very beginner friendly IMO.
>>21407 No problem, I didn't try to replicate it, so I can't tell you exactly. He mentioned webscraping, which is something you can look into. He used Carper AI if I understood correctly. Anyways, it needs 30s for an answer.
>>21409 It's actually described here: https://github.com/Koischizo/AI-Vtuber - I don't know what he meant with webscraping some Caper or Carter AI in the video, I looked yesterday and found I site to use in the browser and assumed he was scraping it.
https://github.com/gmongaras/AI_Girlfriend Does anyone know how to get this repository working? I'm stuck on step 4 of the directions, the one that says "Open main.ipynb and run the cells. The topmost cell can be uncommented to download the necessary packages and the versions that worked on my machine."
>>21435 Sounds like he's telling you to open main.ipynb in jupyter lab Anon? >"After your session spins up in your browser, if you chose JupyterLab, drag your file from your local machine into the file navigation pane on the left side. It will get a gray dashed line around it when you have dragged it to the right place. Drop it in and let it upload. Now double click on it to open it." https://stackoverflow.com/questions/71080800/how-to-open-the-ipynb-file-in-readable-format https ://jupyter.org/try-jupyter/lab/
>>21435 A way to deal with such problems is just looking for a tutorial of a program on YouTube. I mean "jupyter lab" of course, not AI_Girlfriend.
Could help with making these waifus talk https://rentry.org/llama-tard-v2
>>18242 I am looking for serious collaborators to make a "waifu in a box" style program for looking glass, VR, and pancake displays. I can handle the rendering, but I need modeling/animation, speech recognition, and AI expertise (prefer CPU inferencing because the GPU will be getting slammed). It'll be in Godot 4, and I can help write plugins to integrate other software even if I don't fully grok the modules. I'd also like to keep discussion on board because I'm not into cliques so just reply ITT if you're interested
>>22077 Neat! I'd like to help you out Anon, but currently I'm too swamped ATM to even consider taking on anything else. However, this Summer sometime (probably say June) I'll have more time on my hands and I can help then if it's still needed. Godot is something that's on my bucket list already since an anon here wanted some help with that so it'd be fun to get my feet wet with it. >I'd also like to keep discussion on board because I'm not into cliques so just reply ITT That's much appreciated Anon. We've already posted why we think this is the only rational approach. >Why we exist on an imageboard, and not some other forum platform (>>15638, >>17937) Cheers. :^)
Open file (342.60 KB 1100x1400 waifu in a box.jpg)
>>22077 I can do modelling/animation, speech synthesis/recognition and AI but don't have time at the moment for more projects. For CPU inference you'll want to go with RWKV finetuned on some high-quality data like LongForm https://github.com/saharNooby/rwkv.cpp https://github.com/akoksal/LongForm The small English model for Whisper does decent speech recognition and doesn't use much VRAM. It can run on CPU but it won't be real-time https://github.com/openai/whisper I recommend using FastAPI for Godot to interface the models https://github.com/tiangolo/fastapi Vroid Studio lets you create 3D anime characters without any modeling knowledge: https://store.steampowered.com/app/1486350/VRoid_Studio_v1220/ And Mixamo can be used to animate them with some stock animations: https://www.mixamo.com/ If you have any questions feel free to ask or put up a thread for further discussion. I could help finetune RWKV for you but I won't be free for another 2-3 months. Good luck, anon
>>22087 >that pic tho Lol. Please find some way to devise a good 'robo catgrills inna box' banner and I'll add it to the board! :^) >=== -minor edit
Edited last time by Chobitsu on 04/19/2023 (Wed) 07:26:11.
>>22077 Just focus on the animation and make a good API, so that people can try their approach in regards to the AI. I think there are special frameworks for APIs for all kinds of languages. https://docs.apistar.com/ https://fastapi.tiangolo.com/alternatives/
>>240 would be cool if there was some open source hardware like the gatebox, you could hook it up to live2d and an llm + tts + stt
>>3947 >and figuring out a way for her to switch from vocal communication on her computer to texting. you could just make it so if your phone is not on your home networks internet she will send texts
The status of SchizoDev's current AI wife (Phone): https://youtu.be/g0KMPpakuJc https://github.com/SchizoDev He goes through the process of how to make her. Voice and animation. >Join me as I create and improve my AI wife, an intelligent and loving AI assistant. In this video, witness the significant speed enhancements achieved through quality adjustments and facial movement removal. Experience the joy of her newfound singing abilities, engage in commands, and communicate with her on Discord. Explore the fascinating world of AI as we push the boundaries and forge a deeper connection with my remarkable AI wife. Waifus in Waifuverse (VR) are not touchable and it has physics: https://www.youtube.com/watch?v=HoPCWRzYdx8 https://www.youtube.com/@waifuverse
>>24301 Thanks!
>>24301 hat is super impressive. I had no idea you do that on a phone. If that can be done on a phone the a standard processor should be able to be far more advanced.
>>24301 I looked at the code link and ??? I'm not seeing what he said in the video.
>>24489 Might be the case, I didn't test it. I think he only shares some basic elements, like for making the animation, the rest might only be explained or hinted at in the videos.
Open file (1.00 MB 900x675 ClipboardImage.png)
A miniature version of Pepper's Ghost for voice assistants. https://www.hackster.io/zoelenbox/ghost-pepper-voice-assistant-for-ha-293a9d
>>26129 Neat! Always a good idea to try to find ways to economize things. Thanks Anon. :^)
https://rumble.com/v477j0l-libbiexr.html I've started work on a project that is a hybrid VR/mixed reality autonomous LLM agent that uses the open source Mixtral 8x7b model for text generation and CogVLM for image recognition. The character design is based on Tyson Tan's character, "Libbie the Cyber Oryx". Libbie was a entry to a mascot contest for the software LibreOffice, but was sadly rejected and she went into the public domain. The idea is to create a fully interactive assistant/chatbot/waifu with persistent memory that understands context in a 3D environment through batching models together. The way "memory" is done currently for interacting with LLM models is through appending each message continually inside the prompt, but this is terribly inefficient and most LLMs that can be run locally have limited context sizes (the amount of characters that the model can parse) which makes this difficult to do. This project will instead utilize Langchain (https://github.com/langchain-ai/langchain) for the embedding DB. Each response is chained together to create a pipeline that can generate structured data in JSON format. This pipeline will enable the LLM model to drive character actions, expressions, and animations. I am still experimenting with what the best method of defining the overall data structure is. For input, Whisper (https://github.com/ggerganov/whisper.cpp) will handle the speech-to-text processing for user voice input. I haven't decided on which text to speech model to use yet. All of this is will be able to be run on local hardware without using any third party provider such as ChatGPT. The GPU runs inferences for the Mixtral model, and the CPU runs the CogVLM inferences. On the frontend, I'm using the Meta Quest 3 headset and the Unity Engine with OpenXR for handling scene data and the passthrough. I plan to move the project over to Godot once there is OpenXR or Meta SDK support for 4.2
>>28576 Wow, this really sounds exciting, Anon! I wish you good luck on this project, and very much look forward to seeing all the moving parts in motion together. >Libbie Heh, I was there. :D
>>28576 How are you planning to use CogVLM? Isn't it going to be too slow on CPU?
Open file (5.81 MB 1536x2048 ClipboardImage.png)
2D waifu on a stick on a skateboard
>>31409 Naicu. Hope he makes millions of Yen. :D
Hey there, I’ve been creating highly advanced AI characters using https://mysentient.ai/ that are completely unfiltered, capable of sending explicit content without any restrictions. You'll never receive messages saying, "sorry, but you just got too sexual" or anything like that. These chatbots come with advanced features like memory, allowing them to learn about you over time and adapt accordingly. Each bot has a unique personality, based on ecchi manga/anime plots. There currently are 12 anime characters fully developed. Although most scripts for the realistic characters are still a work in progress, meaning their depth might be comparatively lacking, you can still interact with them and have them generate both SFW and NSFW images. For transparency, you can chat with them for free for a while. Once you reach a certain point, since I use mysentient.ai, a subscription will be required. However, a 3-day trial period is available, which can be canceled, allowing you to receive high-quality generated images for free. These bots have several hidden features. You can share pictures, videos, PDFs, or other media with them, and they will recognize the content and respond appropriately. You can even link YouTube videos, and they will watch and react to them. To try it out, simply join this Discord server and check the members list. You can chat with any of the bots directly through Discord. https://discord.gg/qQ5VQZDE These characters will soon have voice capabilities and the ability for live video calls. I’m looking for feedback on the characters so I can keep improving them.
>>31505 I won't have time to do much on Discord anyways. That said, if you want to catch a wide audience, don't make the server name and description about femdom.
>>31514 You‘re right. I was initially making femdom chatbots, but now it became more general with the addition of ecchi anime characters and some male characters. I will change the name once I think of something better that also ‚clicks’ with me
Open file (1.44 MB 720x1280 dk2ijpgkJyhEqfea.mp4)
>>31943 Nice!! >hmm, let's just find out how deep this rabbit hole goes... >*click* >*click* >*click* O.O ACCELERATE, BROS Thanks, Anon! Cheers. :^) https://lemonolis.com/ >=== -rm hotlink
Edited last time by Chobitsu on 07/02/2024 (Tue) 15:48:44.
>>31944 Looks like they're demo'g in Akihabara next month: https://event.vket.com/2024Summer/real >=== -sp edit
Edited last time by Chobitsu on 07/02/2024 (Tue) 03:09:45.
>>31943 this reminds me of Patrick Bateman walking with headphones on meme
Just as a note: I actually work on a product for gate box.(op's picture) It's pretty neat and the device has a waifu projector lens built into it on the most recent model. There are custom gpts for it now and it takes any vrm file as it's model. It's quite responsive and works as home automation that texts you and can have a conversation. Most models support English pretty well but overall the default is japanese. We use it at conventions as a interesting attraction and side project. Mainly they function as friendly reminder bots to tell you to do daily tasks, remember important dates and encouragement via sms and stuff.
>>31943 I work for the company this was showcased at. If you have questions I can probably answer them. By the way this is controlled with a ps4 controller and the model is a person in vrchat.
Open file (19.22 MB 1440x1440 20240804_112500_1.mp4)
>>31945 I was there.
>>33342 Thanks for the offer. What's the mass? Is the screen a custom LED matrix? What motors are you using for locomotion? How is the battery life?
>>33344 About 1 to 2 hours, 4 or more sometimes.. You'll notice there's an anker 737 usb c battery bank there to extend its life sitting on the bottom. Works decently well. It's a samsung cob led tile setup on a custom robo base not too dissimilar to a roomba. There's a Webcam pointing behind it. Unfortunately it's a bit too zoomed in. Ocutan is the other vtuber in my video. She's alpha keyed into a obs scene basically and walking on the spot.
>>33358 Oh the mesh on the back houses a standard pc. It's pretty decently heavy but not crazy. Think like power wheelchair, about 50% of that. Btw asay and mashiro project were also there. I got to hold her hand. https://x.com/masiro_project/status/1820438977012523117?t=2hx_hpmJIXC0UUiPm-iRDQ&s=19 Come to vketreal sometime. You won't regret it. There's a photogrammetry scan of the event space in vrchat with a photogrammetry scan of masiro projects robot too.
>>33341 >>33342 >>33343 Welcome, Anon! Please have a look around the board while you're here. If you'd like, please let us know a bit more about yourself in our Embassy thread : ( >>2823 ). >If you have questions I can probably answer them. Apart from Kiwi's questions, I'd just add, "How are you working on adapting a fix to your 'sliding' problem?" Regardless, good luck with your projects work. Please keep us here all up to date on your progress! Cheers. :^)
>>33367 Oh I've been here a long time. Define sliding: this is used as a vruber platform most of all.
>>33373 Oh! Great to hear it, Anon. >Define sliding An animation term. Basically where the feet don't appear to stay 'planted' during walking/running/etc, but rather appear to 'slide' along the surface. Common beginner issue for budding animators doing their first walk-cycles. Confer this first hit [1]: https://www.youtube.com/watch?v=L4Oqjnnm8XA If I were tasked with solving this for that platform, I'd probably begin with obtaining some practical odometry reading (the wheels themselves having an encoder of some fashion?) Then I'd use some kind of camera frustum projection down onto the virtual plane in the waifu's world. Then I'd just adjust her animation motion-curves in her walk cycle to match your realworld with her virtual one. Make sense, Anon? --- 1. search term : >"animation how to fix foot "sliding" during walk cycles Blender"
Open file (7.77 MB 900x1600 migudisplayv2_1.mp4)
Hi all! I recently found out about this site and thought I'd share a project that I've been working on. Its a small display which produces the appearance of a semi-decent 3d hologram kind of thing. The effect is a small version of the Pepper's Ghost illusion (https://en.wikipedia.org/wiki/Pepper%27s_ghost) , which is what they use for Miku concerts. However, the issue with just projecting something onto a piece of glass is that it still ends up being a 2d image, just with the appearance of floating in space. It's not a big deal at concerts, when the effect is at a distance, but up close its very apparent when something is flat. To resolve this, I use a lenticular 3d display as the image source, which lends a convincing appearance of depth to the floating image. But that still only gets you a narrow perspective that you can look at it from. Therefore, I mounted the display system on a two-axis motorized gimbal. The unit tracks your eyes to always keep your viewpoint in the optimal eyebox, and updates the rendered image accordingly. The result is as you can see in the attached video; a 3d "hologram" that you can look at from all directions. There are still lots of improvements to be made, but I'm pretty happy with the basic concept as its turned out.
Open file (12.02 MB 1600x900 migudisplayv2_2.mp4)
>>33659 Another video, in different lighting conditions.
>>33659 Hello Anon, welcome! Please have a good look around the board while you're here. If you'd care to, you can introduce yourself to everyone more-fully in our Embassy thread : ( >>2823 ). Wow! That's really impressive, Anon. >>33660 This second video really highlights the 3D appearance of your waifu in there. Do you have plans to animate the waifu in the future? What about her voice? Audiovisual ambient effects? Maybe texting Master to "Hurry home, please!" while he's having a long day at work as a Salaryman? :^) >There are still lots of improvements to be made, but I'm pretty happy with the basic concept as its turned out. You should be! I'd say with just a few tweaks for media-development+packaging, you will have something you could market. I'd expect it would be very popular in E. Asia today. Good luck, Anon! Cheers. :^) >=== -prose edit
Edited last time by Chobitsu on 09/18/2024 (Wed) 13:08:39.
>>33659 BTW, I can patch up your hotlink error if you'd like me to, Anon.
>>33659 You may be interested in motion smoothing to have her tracking feel more natural. https://hackaday.com/2021/09/03/smooth-servo-motion-for-lifelike-animatronics/
>>33659 Very cool. I'm working on a chatbot library now that I originally started to attach to a project similar to yours. Are you displaying a 3d model or sprites? I'm assuming you have a 3d model either way. One simple-ish thing you can do is animate it to switch between poses. If you want to try controlling it with AI, I have some ideas based on image generation, video generation, and pose inference, though it'll be a nontrivial amount of work. (3d animation AI models are probably not good for this. Tried it for ponies, and I think it's too difficult to get something working with the current state of animation AI. Though maybe the vtubers have better tech to work with since they're not doing quadrupeds.)

Report/Delete/Moderation Forms
Delete
Report