Unboxed the new laptop! Yay! Tried installing Fedora 15 x86_64 from a live USB stick - no dice. The thing didn't even boot, giving weird intel MUX INFO errors and falling to a debug shell (dracut). Tried installing Ubuntu 11.04, and that just crashed. Took it to Pras to try and identify the issue, and eventually realised that the live USB image was at fault, and the OS got installed fine using a burnt disc. The video drivers look buggy, but it can be worked on. Wifi seems to be an issue - had to install the specific uname file using yum. iwl6000g2b-firmware-126.96.36.199-2.fc15.noarch Otherwise I had to build the compat-wireless package, which this install avoided. Had to manually configure the wifi using the terminal the first few times using iwconfig wlan0 essid "CMU" and then playing around with the wifi applet Had a few issues with installing Adobe flash player, and eventually used the 64 bit square player and copied it to the mozilla/plugins directory, and it works now Tried installing ROS using the instructions. It didn't (as usual) compile nicely. Had to install boost libraries, but it still showed an error on make - specifically an example program didn't compile, and threw an error stating an undefined "boost::filesystem3::path::extension() const" A bit of googling (http://answers.ros.org/question/996/boost-symbols-not-found) made me realize that this was an issue with the changeover from boost::filesystem from v2 to v3, and so I had to add a #define BOOST_FILESYSTEM_VERSION 2 to all files including a #include <boost/filesystem.hpp>. To find these files, I used find . | xargs grep 'boost/filesystem.hpp' -sl this gave me about 9 files which I modified, and ran the script again. Success!
Tried installing Eclipse, the dump crashed whenever I tried to select preferences, Fixed it by using the repository version of eclipse - worked like a peach - just had to add the helios packages for dependency resolution from within the preferences Installed NetBeans 6.9 from the repo
Worked on ROS tutorials. However rxplot did not run, giving a missing ImportError: No module named backend_wxagg error. Installed matplotlib and all wx packages and rosdep install rx, but to no avail. Seemed like a dependency issue. Used the following command yum install python-matplotlib-wx to get the missing dependencies from (http://groups.google.com/group/sofastatistics/browse_thread/thread/e41f51d73349048c?fwc=1&pli=1) The rxplot has weird synchronisation issues the plot jumps at 1 Hz and eventually messes up the entire x and y plot. The echo messages, when plotted do not show this behaviour. BUG, most definitely
Went through ROS subscriber and message and server client examples Attended the Airboat Meeting 2 Airboat Meeting 2 Discussed image parameters, and lens test performed by Louis. Decided that the fish eye actually gave a much wider field of vision vis a vis the regular stock camera without sacrificing too much resolution. The problem of reflection still remains, and adding a polaroid filter might add further distortions. The camera can most probably not be set in a prone position, and so it will be left standing (As of now) Wrt the mechanical part the new foam pour has (perhaps) finally been decided, and robot city needs to be prepped up by next week - upto Captain Abhinav :) Balajee most probably won't be available till end of July. Discussed manoeuvering issues with Pras, but he seemed confident of the boat sailing through, especiallyas the closed loop worked pretty much perfectly on Airboat 1. // Discovered the University Centre Basement - It has an awesome 'Scotland Yard Room' which has a ping pong table, pool tables, soccer tables, a pinball arcade and a dance machine. Had fun playing on it with Remus, Mihai and Piotr
Installed rosjava all over again - this time had more directories to work with. in ~/ros/rosjava/android went into tutorials and library, and modified the build_projects.sh to point to android-10 instead of (weirdly) android-9 now opened one tutorial project in Eclipse - the standard Android app way. Library paths were messed up, so deleted them, and included them from rosjava/lib Library (and project) issues - used File->Import existing project to include ~/ros/rosjava/android/library and ~/ros/rosjava/rosjava Issues decreased significantly 127.0.0.1 Need to set nodeConfiguration.setHost..... and set rosCore.createPublic So, by statically assigning the nodes to run on the device local IP address, and the roscore master to be running on a nother machine in the network, we can get ros working :D roscore running on phone is not a full implementation as yet - cant interact with it yet - can only see published things --- Also attended a MAV meeting where Sebastian presented a draft of his presentation on river mapping and navigation on MAVs. He uses a pretty decent approach to find out what region in river, and uses LIDAR to scan the banks. He tracks the path plotted by taking into account the GPS, IMU and Optical Correlation through a Globabl filter. The Optical Correlation shows a significant drift, which is corrected by the other two. The processing is done on board with 4 ARM dual core processors in parallel with a DSP
Started working on establishing message framework for rosjava implementation. Filled the board with the decision alternatives. Found out that the data transfer between nodes is NOT done by XML RPC - XML RPC is used just to establish connection between the two - the data is serialized and sent over using in built methods over TCP or UDP Started Eclipse, to get weird Java compiler issues that it required .class compatibility set to 5.0. Seems that something, somerwhere messed with the default settings, so had to manually set the Java compiler version in Project properties to 1.6 The code compiled then. (Probably because of the running JDownloader?) Observation: Changing screen orientation resets the code. Probably because it onPause()s the app on changing orientation. Discussed the implementation of the Airboat server with Pras. So the way forward is to create a barebone implementation of an interface (preferably) or an abstract class library of rosjava precompiled within it, so that this library could be implemented in any java project. Specifically, the implementation of this library for a real airboat will declare how, say, setWaypoint() works by selecting a controller and configuring it using PIDs. However, for a simulation, these methods could be declared so as to only move in a cartesian space on the GUI, or however the simulator wants to move the boats. Decided which topics were to be selected as topics that are broadcast and which topics are to be implemented as services. Not everything needs to be published. The TCP or UDP implementation is done by ROS on its own. Higher order functions like setAOI(Area Of Interest), setNewThresh(To set the 'newness' threshold for the image) and airboatStates (Which stores the state of the other boats) are not crucial to the core functionality and can be implemented with the boat control implementation of this bare bones library.
Tried running the different tutorials on the android app. Figured out that with roscore working on my laptop the service on the phone couldn't communicate due to the firewall. Watching for topics on the syrah server worked, but no image was forthcoming when I used the image_transport tutorial. Will figure that out soon. While trying to show this, realised that I needed to rosmake image_view. This (almost immediately) threw errors (how predictable) Installed jasper, ffmpeg and GraphicsMagick -devels and managed to rosmake libplugins. Also had to install libuuid-devel Howeverthe build started failing with very similar errors to those encountered on June 21 with the boost libraries used in the ROS code being deprecated. Performed the same manouever by editing the boost_fs_wrapper.cpp Old friend OpenCV2 then started playing up - it threw a whole host of ptrdiff_t errors. Fixed it by editing build/opencv-svn/modules/core/....core.hpp and including <cstderr.h> Voila! Build complete! :) sudo yum groupinstall "Development Tools" "Legacy Software Development"
Finally managed to view the Android camera using one of the tutorials in /android/ It was pretty cool. Realised that image_view required a few dependencies that hadn't been compiled - theora and compressed. Also, the code did not follow the standard ROS convention of the compressed image being published to camera/compressed and published it to /annonymous/camera instead. Hence fixed this to publish it to /camera/compressed and voila, the image could be viewed from any node. Now all that is required is to implement this with actionlib
Attended the airboat meeting where we discussed results of the imagery tests performed by Pototo using the curved mirrors. Also discussed the mechanincal component of design with Chris and Abhinav. The foam material seems to have been finally chosen, and the acrylic base of the prop needs better bearing support as it often interferes with the makeshift channel that Chris carved out for it. Managed to implement the fibonacci actionlib example by running the actionlib server on the android app and listening to the broadcasts using the stock SimpleFibonacciSequenceClient app after pointing it to this server. All that needs to be done now is to link the airboatServer components in it.
Worked on the
Airboat Meeting 3 Discussed specifications for the test to be conducted at the irrigation lake near Baltimore Primarily, the plan is to perform sensor tests - pH, conductivity and temperature over the 120x60 m pond. The plan of action is to decide the lake into a grid of 5x5m sample patches, and proceed with further navigation - three methods have to be implemented 1. A Random Walk/ Lawn mower 2. Prefer Highest std. Dev./Err. 3. Prefer high gradients So the work to be done is Paul: Autonomy setup Abhin: Sensor Calibration KSS: Control + Image connection Louis: Bank avoidance Also discussed about the requirements of the boat to stop and possibly rotate to point towards a specific object or direction (specially for capturing imagery) Discussed about the depth of the sensors required, and if a retractable sensor rod could be used. Camera position has to be decided, still. Pras plans to have a quad on the boat as well, so might occlude the view. However putting up a phone that way up poses issues.
Finally started working on actionlib message generation - realised that I needed to generate a .action file. So used a pose for Goal, and two Vector3s, etc. as per the requirements. Didn't implement the Camera straightaway - that's deferred for later Generating the messages was harder than expected. Trying to use the .action in own project just didn't generate anything! Anyway, copied it to actionlib_tutorials, but only to be greeted with compiler errors stating that package org.ros.message.geometry_msgs does not exist was not available :S A better look into things leads me to the conclusion that the python script may not be at fault in this respect, and that it does not need to be modified. However I had to modify a few source files. The issue that occurs is that during compilation of the roscpp.jar file, issues are generated. Something like this [javac] home/kshaurya/.ros/rosjava/gen/srv/roscpp/org/ros/service/roscpp/Empty.java:4: package org.ros.internal.message does not exist [javac] import org.ros.internal.message.Service; Eventually I realized that the test_ros messages were not being generated, and so I had to manually generate the message and service files to ~/.ros/rosjava using the python scripts, change their path to point to /~.ros/rosjaava/gen/org..... AND modify the three roscpp files (Empty, GetLoggers, SetLoggerLevel) to import org.ros.message.Service rather than org.ros.service.Service. Fixed these to get things working again. In addition Pras has added a crw-cmu thing too, so I guess things are set now. With regards to the architecture, multiple actionlib servers shall be set up - one for say pure waypoint navigation, another for camera, etc.
Managed to mess up wifi earlier in the day by uninstalling NetworkManager. Had to reinstall it using a rpm downloaded off a separate computer. Finally started working on the actionlib implementation. Created a simple .action file for waypoint navigation, and included it in actionlib_tutorials. Regenerated the jar files, and included it in the crw-cmu project. Also added geometry_msgs, etc. Mostly copied the fibonacci example, omitting unnecessary stuff. Got access to the repository (yay!) and pushed my commits. Pras later fixed some of the stuff and the autogeneration code.
Played around with the callback functions. The sac.watForResult method doesn't seem to work that well (perhaps there's some issue in its implementation) Talked to Pras, and realised that ActionLib was to be used for setting the *configurations* and not for accessing/polling data. So the way it works now is that the VehicleServer interacts with the outside world via actionlib for config messages such as set camera on at so and so params, stop taking pictures, go to so and so waypoint, stuff like that. The publishers take care of sending out the data streams as such.
Implemented a dummy waypoint controller using actionlib. I shifted the separate RunActionServer and RunActionClient into a single SimpleBoatSimulator class as launchServer and launchClient methods. Committed the code. Pras said that we could possibly get this stuff now working on the boats in a couple or so days.
Sat down with Pras to set up the Boiler plate code to encapsulate all the wrapper classes into a single class.
Fixed the bootup issue on the laptop by disabling the sendmessage service. Bootup times dropped from ~3 mins to 13 seconds! XD Also fixed the RosServer calls for multiple actionlib instances. The issue was that each actionlib instance was using the same /status, /goal, /result ,etc. topics as they were assigned to the default (root) namespace. To fix this created a separate Node configuration for each of the the buildSimpleActionServer instances and assigned them a different node namespace by using the NameResolver class
Airboat Meeting 4
Saw the latest working prototype. Decided on the date of the visit to the irrigation pond. George was present too, and he mentioned some new low cost small form factor integrated sensor that a company was intending to launch soon. Also decided that shelling about $70 more would get us new Google Nexus Ss which had a 3 axis gyro integrated, which saves the trouble of purchasing an external IMU at the same cost, with added integration issues. George said that the sensors that the scientist at the irrigation pond uses are attached to buoys shaped like cowboy hats. The water recirculates through all the fields, and hence they really need to measure the water quality at all times to ensure that no pollutants (or pathogens) enter the water stream. There was also talk about some BOD thing, don't really remember what it was exactly.
Dental, Aircraft assembly haptic.buaa.edu.cn Devices : Omega, hantom Desktop, Haption 6D, Delta Can be serial or parallel Back drivability, singularity problem - require a task space without a singularity (Gosselin, 90) ax + B Theta = 0 Internal singularity => Det A = 0 To evaluate how far away from singularity, Gosselin, Pottmann, Kozak - a physical representation, rather than a mechanical one Voglewede, et al So the deal is to narrow down the singular subspaces to those created by the Force applied and those by the moments. Once the two subspaces are obtained, the complement of their union is the 'safe' workspace to implement. (the union is called the Pseudo singular space) To determine these parameters Mf and Mm, we use Rayleigh quotients So, there were two hinged joints on each side of the trainguular base - spherical joints 4 bar linkage to give a 6 DoF end manipulator - movement and twist Hao Li, Yuri Zhang, IEEE 2011.5 No issue with low frequency human jitters ~7Hz. Devices are working at 1 kHz....:S ------------------- Umbrella Based Texture Bombing Create carpetted textures - mimic diversity with limited seed vals Related works - texture bombing, Diversification, Level of Detail, Create a polygon from a characteristic shape to create an umbrella, and then modify the shape to create modified umbrellas by morphing the end point s of the polygons generated - and the colour can be changed as well - so morphological diversification Divide the space into grids, and map the umbreallas such that atheast a grid square overlaps (mipmapping for places where high LoD is not required) So for things further away, you need to represent it by lesser object if possible To do this they cluster using K Means, find the convex hull, and merge the overlapping umbrellas, blending their individual colours This works as umbrella footprint is really small To make things more realistic, add shadows using pose information, and lighting using knowledge of global lighting. Allows smooth rendering in real time. Brain Computer interface thesis Samuel Clanton Kleins motion control (impedance velocity controller) 7 D decoding model arrays inside cortex Spike rate = delta V / time R,P,Yaw not vectors per say, so angular velocities used 6+1D how to calibrate decoding model? a. Use the model on monkeys arm, and then use that b. Use observation data - the monkey saw humans do it, or it tried random things till it figured out what to do no better hemisphere for control Impedance Vellocity controller Kleins controller - spring+damper attached robot hand However the monkey doesnt always think about moving the hand so an active shared control system used - mixing operator command and auto command which gave perfect commands as per on robot configuration LAter used passive - which attenuates portions not positively spanned so these work as guiding mechanisms So to get the monkey to learn very well the error admittance is gradually reduced to make it less and less guided The monkey started using the force feedback to align the robotic hand quite often instead of actually trying to rotate it to orient itself to the target the feedback is visual.
Airboat Meeting 6
Tried to analyze the spinning in the software crash Taking three boats, implementing two. Couple of extra hulls. Electronics for three boats The EC measurement is instantaneous, but temp and oxygen might have transience Getting the boat to drivve around and GET measurements is omre important as an issue Need to perform range tests An autonomy toggle will be required to cross the linie between the buoys and resume, so the boat just realises that it has 'jumped' to the new pos - no need to restart Pras talked about calibration. Need to figure out the horizon line deal, even if its constant.
ToDo 1. ROS Tutorials 2. Make rosjava 3. rosjava tutorials 4. Need to figure out messages in ROS for interfacing 4.1 Try out ActionLib 5. Pure Java library to blackbox all the ROS stuff - a wrapper of sorts 5.1 Fix repository layout, and use crwlib jar 5.2 Work on Image transport code 5.3 Work on figuring out RosServices July 25
Had issues with the wifi again. Hopefully fixed it by lowering the frequency using iwconfig to 2.462 MHz (I think thats g speed) Pras fleshed out the interfaces. A Test package has been created to test out stuff. A few messages were changed, and so had to delete archaic entries. Tried to visualise stuff using rviz. Realised that would
Had to make flyers for the RISS thing Talked to Paul, and started working on implementing VehicleServer for his code
432254.86, 4371539.65, 0.00 18 north 432254.86, 4371539.65, 0.00 18 north base station 432227.44 4371540.67 launch 432206.243 4371541 Right 432246.1 4371543 Right bank, green bush 432248 4371589 432248 4371544
Testing in schenley It is heading correcctly, but significant GPS drift Goal is sent repeatedly ad infinitum. Couldn't see completed on the GUI, whereas the controller DID stop on reaching the waypoint. Have to clcear curret waypoint on achieving completed flag. VERY SIGNIFICANT GPS drift. Phone had to be yawed perpendicular to the box axis
Talked to Paul about further work this week. He mentioned that I could work on one of two things- the controls or the IP based suff. So I decided to do something on the image queing methodology. Paul said that in doing that I could possibly economise on the images that I send, for example sending images only when the boat changed position significantly - so for example a yaw, or after moving say 10 metres - he would then implement an interface for that, and that we could write a paper/something like that that on that. Airboat Meeting x Hulls Go with three copies of the best design Shrouds similar shrouds, with minor tweaks mesh Temperature sensor at the hardware level add it as a sensor channel Going to try a heatsink attached to the ESC before doing anything else Performance eval criteria --- Contact info on the boat Behaviour for comm losses - WHEN to turn around... Hence easiest way is getting a go home app, which overrirdes the existing setup Restart remotely Talk to brad for obstacle avoidance Queue up packets in spotty comm Add connectivity parameters in sending in data (good connectivity => send more)
Observations from discussions with Pras on the Tests
1. Weird Network bug, when topics did not show up on my roscore, but did on connecting Abhinav's - very probable that it's a network bug 2. Possible boat interactions - How to broadcast current state to other boats 3. How to implement novelty basis for taking pictures that are situationally aware - first elemental step would be to take pictures when the boat changes it orientation or position. More intersting ways to go about it will possibly using a. A geometrical based pie sector based mapping b. A grid based mapping where the controller tries to take a path to goal that will actually acquire the maximum number of images along the path. Such a navigation methodology differs from a path finding methodology in that the future path relies on the path previously taken to reach a particular point - you really dont want to keep moving in a straight/shortest path in which you, say, keep looking at the goal all the time. 4. A debug button to start a new log every time the operator wants one during deployment 5. Integrating Pototo's code - consider the image and the obstacles as a pseudo laser scan, with distances from the base to the FIRST obstacle converted to metric distances and published to a controller. We could publish to cmd_velocity, with an autonomy controller sitting around pretty, and an external publisher doing the dirty stuff for the mean time. 6. 7.