Next version 'Nixie' is coming soon !

Help us to get ready with MyRobotLab version Nixie: try it !
Javadoc link

Update 2013.05.
 

 

Here is the effect of fluorescent lighting, the waves I suspect are caused by the constructive/destructive 60 hz cycles.

Javadoc link
Example code (from branch develop):
#file : Arm.py (github)
# start the service
arm = runtime.start("arm","Arm")

Arm needs development.  The proposal is to have a service where lengths, joints and limits can be specified and forward or inverse kinematics can be used to predict or plan spatial location.

Javadoc link
Example code (from branch develop):
#file : Pid.py (github)
#########################################
# Pid.py
# categories: pid
# more info @: http://myrobotlab.org/service/Pid
#########################################
 
import time
import random
 
pid = runtime.start("test", "Pid")
 
pid.setPid("pan", 1.0, 0.1, 0.0)
pid.setSetpoint("pan", 320)
 
pid.setPid("tilt", 1.0, 0.1, 0.0)
pid.setSetpoint("tilt", 240)
 
for i in range(0,200):
 
    i = random.randint(200,440)
    pid.compute("pan", i)
    
    i = random.randint(200,440)
    pid.compute("tilt", i)
    
    time.sleep(0.1)
Example configuration (from branch develop):
#file : Pid.py (github)
!!org.myrobotlab.service.config.PidConfig
data: {
  }
listeners: null
peers: null
type: Pid

A PID service allows the tracking and transformation of output in relation to iinput.  This is currently used as one of the tracking strategies in the Tracking service.  Input is sent to PID, a "compute" method is sent and appropriate output is sent to a servo.

References

My purpose is to just become more familiar with MRL.

I was trying to replicate the sample tutorial that you have and seem to be having some trouble.

I am interested in getting the speech aspect working for a simple project of face recognition and a comment of recognition.

as for right now though I would just like to get the example working.

Here is a screen shot of what I have done.

Yay ! - LKOptical tracking filter tracking 30 points simultaneously !   The Tracking service can proceed .... Wooh .. about time !  Still need to write a bug up for JavaCV, but the filter works around it.

Added an ouput panel - which allows recording of a video stream or individual frames from the output stream.

Javadoc link
Example code (from branch develop):
#########################################
# ThingSpeak.py
# more info @: http://myrobotlab.org/service/ThingSpeak
#########################################
 
# virtual=1
comPort = "COM12"
# start optional virtual arduino service, used for internal test
if ('virtual' in globals() and virtual):
    virtualArduino = runtime.start("virtualArduino", "VirtualArduino")
    virtualArduino.connect(comPort)
# end used for internal test 
 
readAnalogPin = 15
 
arduino = runtime.start("arduino","Arduino")
thing = runtime.start("thing","ThingSpeak")
 
arduino.setBoardMega() # setBoardUne | setBoardNano
arduino.connect(comPort)
sleep(1)
# update the gui with configuration changes
arduino.broadcastState()
 
thing.setWriteKey("AO4DMKQZY4RLWNNU")
 
# start the analog pin sample to display
# in the oscope
 
# decrease the sample rate so queues won't overrun
# arduino.setSampleRate(8000)
 
def publishPin(pins):
  for pin in range(0, len(pins)):
    thing.update(pins[pin].value)
 
arduino.addListener("publishPinArray",python.getName(),"publishPin")
 
arduino.enablePin(readAnalogPin,1)
Example configuration (from branch develop):
!!org.myrobotlab.service.config.ServiceConfig
listeners: null
peers: null
type: ThingSpeak