as a matter of fact I am. I’m using the sample software provided by Amazon, and adapt it to be able to talk to the mirror.
It has plenty of advantages :
- not using AlexaPi, which is a third party software module, possible containing extra bugs (with all due respect to the developers of course ; respect), where as I’m supposing the Amazon samples all run pretty good
- Amazon supports it, and when new features arrive, new samples would theoretically be easier to integrate without having to code them myself or wait for the alexa pi community.
- All features are supported out of the box: music playback, alerts, feedback cards. These features are not all supported by AlexaPi, I think (or I should be mistaken).
- Wake word detection provided by Amazon, not using sphynx etc (or using a hardware button for that matter). Something provided from the vendor, free of charge (for personal use).
- Automatic integration of custom skills.
- My interface will use a simple communication platform in which practically no external code is required, except for some java code extending the java client provided by Amazon. But that is rather straight forward.
Any questions, tips, pointers, etc. Keep them coming, and subscribe to the blog to keep up to date with my progress. Eventually I’ll be uploading the whole shabang to GitHub, but you can see all the code & the explanations in detail right now on the blog as well.