Wednesday, October 7, 2009

Demo photos

Two demo pictures...

With reMind, you could connect your phone to your favorite toy, and use it to pull out web info for you during meetings...


Or simply discuss what you see in a magazine...


BTW, we've posted full project submission on the googlecode site (link on the link tab)

Friday, October 2, 2009

Windows demo out

I've packed up some win32 demonstration code, in http://code.google.com/p/maemoremind/downloads/list using all of the mentioned open source code and packed up in python+jython. Read the README for installation instructions, but don't expect a convenient install script... (yet)

Thursday, October 1, 2009

First conversation with remind!

After using jython to bind freetts and sphinx 4 (both java libs) to python and running a PyAIML server, I managed to hold this conversation with the machine:
me: Good morning
remind: Hello and how are you this morning?
me: good
remind: Thanks for the compliment
me: how are you?
remind: I am fine, thank you
me: what did you do today?
remind: Ah, I talk to people on the web. What do you do?
me: I am a programmer
remind: I am glad to hear that you are program

Some cheats were involved - the programs ran (in python and java) on my win32 computer, and I hit a key every time I finished a sentence.
Audio File link:
http://maemoremind.googlecode.com/files/remind_talk_01_10_09.wav

Thursday, September 24, 2009

Architecture Plan

So, how will it work?
The project is going to have three parts:
AI part - will be written in AIML - an xml language for creating interactive chat bots. The AIML will be processed by pyAIML, and will be based on the ALICE chatbot AIML files.
Speech part - speech will probably be analysed by CMUSphinx - an open source speech recognition library, in order to generate text. Speech will (probably) be generated, in Java, by FreeTTS.
Modules - some python modules will be written by the project team, in order to enable special functioning (downloading information from web resources, accessing phone functionality such as taking pictures...)

Most programming work will go to:
1. connecting the python server which will run the AI and the modules, and the (probably) Java server which will run the speech recognition and generation.
2. designing modules

What is this?

reMind is a project for the N900Push competition. The concept is to make an N900 cellular phone run a full-blown talking AI - you will be able to go into the room and start talking to your phone. Even better, you will be able to configure multiple personalities on the phone, and bring your household objects (dolls, toys) to life... The technology used will be composed of several open source packages.
The next phase is modules - in some points in the conversation, the bot might go on wikipedia to search for some info, take a picture for you, and all in natural speech, not specific commands ("please take a picture", "photo me", "take a photo now", etc.). Other modules can improve the brain, for example adding the ability to analyze google news articles and talk about them.