Conversation
|
Hi @flov , thanks for your contribution. My billy hangs on the toilet wall, so I didn't want to add a camera myself haha. But it would be a fun optional upgrade ! Could you update your branch by merging my recent changes from main first? I have done some refactoring recently, causing the session.py file to split up. I will definitely try it out, I will order a picamera but in the meantime also test with a usb cam. I think we can just hook it also into the 'normal' realtime session directly as the gpt-realtime models support also image inputs: |
|
haha, yes, having a fish see you while you're on the toilet is indeed a bit creepy 😂 |
|
btw. I've tried to play chess with him with the python-chess library, but it didn't work so well. He kept on making illegal chess moves :D. I think chat gpt is not good at playing chess. I think it would be possible with the integration of stockfish, but it's not that trivial |
Billy can now take a photo and describe what he sees. Adds: - core/camera.py: capture via picamera2 (Pi) or OpenCV (Mac/USB) - look_around tool registered in base_tools and handled in session.py - CAMERA_ENABLED / CAMERA_DEVICE config vars - test/list-cameras.py helper to identify camera device indices - README section M documenting setup and usage
This PR adds camera support for the billy-b-assistant. Right now it can look around and describe what it sees.
GPTARS inspired me to have it play a round of chess with me.
In the future I would like to get people to play a round of chess with me via the camera.
Summary
look_aroundtool that lets Billy take a photo and describe what he sees using GPT-4o-minivision
enforced via system prompt
picamera2), USB webcams, and MacBook FaceTime camera(via OpenCV) for dev/mockfish use
New files
core/camera.py— frame capture (picamera2 → OpenCV fallback) + GPT-4o-mini vision API calltest/list-cameras.py— utility to enumerate camera devices and pick the rightCAMERA_DEVICEindex
Configuration