Original Article A reproducible educational plan to teach mini autonomous race car programming International Journal of Electrical Engineering & Education 2020, Vol. 57(4) 340–360 ! The Author(s) 2020 Article reuse guidelines: sagepub.com/journals-permissions DOI: 10.1177/0020720920907879 journals.sagepub.com/home/ije Süleyman Eken1 , Muhammed Şara2, Yusuf Satılmış2, Münir Karslı2, Muhammet Furkan Tufan2, Houssem Menhour2 and Ahmet Sayar2 Abstract As autonomous cars and complex features of them grow in popularity, ensuring that analyses and capabilities are reproducible and repeatable has taken on importance in education plans too. This paper describes a reproducible research plan on mini autonomous race car programming. This educational plan is designed and implemented as part of a summer internship program at Kocaeli University and it consists of theoretical courses and laboratory assignments. A literate programming approach with the Python language is used for programming the race car. To assess the educational program’s impact on the learning process and to evaluate the acceptance and satisfaction level of students, they answered an electronic questionnaire after finishing the program. According to students’ feedback, the reproducible educational program is useful for learning and consolidating new concepts of mini autonomous car programming. Keywords Reproducible research in robotics, mini autonomous car, robot operating system, deep learning, Jupyter Notebook, electronic questionnaire 1 2 Department of Information Systems Engineering, Kocaeli University, _Izmit, Kocaeli, Turkey Department of Computer Engineering, Kocaeli University, _Izmit, Kocaeli, Turkey Corresponding author: Süleyman Eken, Department of Information Systems Engineering, Kocaeli University, _Izmit 41001, Kocaeli, Turkey. Email: suleyman.eken@kocaeli.edu.tr Eken et al. 341 Introduction Reproducibility is considered an attainable minimum standard for evaluating scientific claims when fully independent replication of a study with different data is not feasible. This standard requires researches to make both their data and computer code available to their peers, but it still fails to achieve full replication, since the same data and not independently collected data are analyzed again. Nevertheless, this standard while limited still allows for assessing the quality of scientific claims by verifying the original data and codes. It managed to fill the gap between full replication and none at all; after all, the reproducibility of a study can lie on a scale of possibilities rather than just the two mentioned extreme points. Reproducible research has been a recurring subject appearing in high-impact journals. The state of reproducibility has been studied thoroughly in some domains, e.g., in computer systems research1 or bioinformatics.2 Some studies have covered the field of robotics research as well. Compared with other fields, robotics is special in that it aims to enable an embodied agent to interact with the world physically. Research on robotics has been concerned with creating efficient models that meet requirements like body dynamics rigidity, or multiple-view geometry that determines an action output based on the perceptual input, or predict the perceptual output based on a certain action. This, while working fine in a controlled setup with well-defined variables, made it a difficult challenge for robots to move out of these environments to the real world with its unpredictability and the wider range of tasks and situations. Robots are significantly and quickly improving regarding how they perceive the world around them, they can now sense the world in new unprecedented ways: see, hear, read, touch, etc.3–5 The ability to deal with these new inputs and learn from them pushes the capabilities of robots steps ahead. Even though researchers have studied and applied various machine learning approaches on this task, this often comes with a caveat represented by some assumptions or simplifications in the mapping between perception and action and vice versa.6–8 This, however, can be alleviated, thanks to the availability of big data and the improved machine learning approaches to benefit from it.9 Unlike computational science disciplines, robotics research faces practical challenges when it comes to replicating or reproducing results for objective assessment and comparison. The difficulty of perfectly controlling environmental variables, poor understanding of comparable evaluation metrics or finding goal similarity across different domains, determining what information is needed to allow result replication, they all contribute to making this challenging. To make matters worse, we are yet to build a solid theoretical foundation for experimental replicability in robotics, hindering research progress and technology transfer. Though it is worth noting that recent years have seen significant movement forward in this regard, with workshops10,11 and summer schools12 covering the design, planning, experimental methodology, and benchmarking of reproducible research in robotics. IEEE Robotics and Automation Magazine also introduced a new type of articles, R-article. It includes a description and access information to data sets and the 342 International Journal of Electrical Engineering & Education 57(4) source code or its executable files on CodeOcean,13 as well as hardware (HW) description and identifier,14 with details on the tests that have been run. Teaching and learning autonomous vehicle programming can be a difficult task, and adding robotic vehicle design and construction to the mix only makes the learning curve even steeper. This paper presents a reproducible research plan to teach autonomy using state-of-the-art sensors and embedded computers for mobile robotics built upon a 1/10-scale racing car platform. A combination of theoretical courses and laboratory assignments has been given to undergraduate students using this methodology as part of a summer internship program at Kocaeli University. First of all, we covered the theoretical courses of this program, this includes subjects such as basics of commanding the vehicle actuators (steering and drive motor), robot operating system (ROS) fundamentals, leveraging data from the Light detection and ranging (Lidar), the Inertial Measurement Unit (IMU) and the wheel odometer, basics of control systems, basics of blob detection in images, fundamentals of robot perception including visual servoing. Then, students have to implement what they learned in the laboratory exercises. The 10 laboratory experiments are designed in a way to provide the students with hands-on experience with the mini autonomous vehicle programming in this research plan. These experiments offer a fitting laboratory solution to bridge the gap between the theoretical knowledge acquired from the courses and the real-time application. The contributions of this paper can be summarized with: • We have presented modern scientific computational environments for programming mini autonomous vehicles using a literate programming approach with the Python programming language due to its popularity within the scientific computing community. The environment makes it possible to combine algorithmic details and explanations with the implementation code in one file. • We have evaluated the reproducible educational plan based on students’ surveys. The rest of this paper is organized as follows: section “Related work” presents related work about courses and robotics competitions with autonomous vehicles and similar platforms for teaching this subject. Section “Used platform” describes the open-source HW and software of the platform. Section “Educational approach” introduces the plan including theoretical lessons and laboratory assignments. Section “Evaluation and Discussion” gives students’ feedback on our reproducible educational program and discussion of the results. Finally, the last section presents the “conclusion” of this work. Related work The use of autonomous vehicles is becoming a common choice as a tool for teaching several related subjects, including robotics, electronics,15 programming, and complex control. Carnegie Mellon’s Robot Autonomy course pioneered this Eken et al. 343 approach in 2002 and set the framework for using creativity to spark the students’ interest in learning.16,17 There is also LEGO MindStorms NXT, a low-cost programmable robotics kit that has seen far reach out for it managed to advance the ability of teachers and learners of all ages to produce autonomous vehicles in an easy and timely manner, thanks to its canned libraries and a powerful graphical programming language.18 Competitions and camps also have shown incredible results, setting them as effective tools in motivating students to extend their knowledge and improve their ability to program a robotic vehicle to run autonomously.19,20 The FIRST Robotics Challenge, the FIRST Lego League,21 BEST Robotics,22 National Robotics Challenge (NRC),23 and EARLY Robotics24 are a few examples. These programs are geared toward precollege students and cover a wide range of choices of robotics HW for students to build and experiment with. There are many other examples25,26 of successful approaches to teaching the subject. During this program, we took advantage of the flexibility and extensibility of the Jupyter notebook web application in our development environment, this approach facilitates both the educational and the reproducible research workflow. In addition to those benefits, the use of the popular Python programming language minimizes the effort required to get up and running with the environment and get a hold of its syntax and semantics. Moreover, the availability of free and opensource packages for this programming language is another reason for choosing it; they can be easily integrated into the environment, extending it to suit whatever the developer needs. The platforms used to teach autonomous control are as varied as the curricula for teaching; this variation covers a wide range of capabilities and price points.27 They also vary in how robust they are, from children toys to industrial equipment passing by hobbyist levels. Most notable examples at the lower price end include the use of Lego kits,21,28,29 kits based on the iRobot Create fitted with either a Gumstix-embedded computer30 or a laptop.31 These platforms serve the important purpose of lowering the entry bar for as many students as possible, especially those with limited budgets. However, the majority of them lack state-of-the-art sensors and computers that power and make many contemporary robotic systems possible. The platform we opted to use on the other hand includes the state-of-the-art HW utilized in real-world robotic vehicles today. Detailed information about the platform is given in the following section. Used platform In this section, we will describe the HW configuration of the car kit and its features. Racecar, or Rapid Autonomous Complex–Environment Competing Ackermann-steering Robot, was initially developed for a competition by MIT in 2015, then updated in 2016 to be used for teaching robotics.32 The kit uses a 1/10 scale Traxxas RC Rally Car racing car as its backbone (see Figure 1(a)), accompanied with VESC,33 an open-source electronic speed controller. And the Nvidia 344 International Journal of Electrical Engineering & Education 57(4) Figure 1. Hardware parts of used car kit. (a) Racecar mini autonomous developer kit (b) and its hardware components. Jetson TX2 developer board as its main processor, which comes with 256 CUDA cores. The following are a list of some advantages associated to open source systems: lesser HW costs, no vendor lock-in, simple license management, abundant support, scaling and consolidating, low-cost and large amounts of available information on the Internet. As for sensors, the kit includes three types: a stereo camera, a Lidar, and an IMU. ZED stereo camera34 allows for the use of standard stereo matching approaches to acquire depth information from two images by two different cameras. The Stereolabs Software Development Kit (SDK) implements a semi-global matching algorithm that works on Graphics Processing Unit (GPU)-based computers, which makes it suitable for our platform with Jetson TX2. 3D mapping Eken et al. 345 Table 1. ZED camera video modes. Video mode Frames per second Output resolution 2.2K 1080p 720p WVGA 15 30 60 100 4416 1242 3840 1080 2560 720 1344 376 also can be done using the same algorithm. The camera is capable of working in the following video modes. The camera is capable of working in the following video modes as tabulated in Table 1. The second sensor is the Scanse Sweep Lidar.35 It can operate at a distance of 40 m away collecting 1000 samples per second. The Sweep has a head rotating clockwise while emitting a laser beam to collect data in a single plane. Finally, we have Sparkfun 9 degrees of freedom (9-DoF) Razor IMU M036 as the IMU. It has a SAMD21 microprocessor paired with an MPU-9250 9-DoF sensor, making it a small and flexible IMU. This means we can program it for various tasks: monitoring and logging motion, transmitting Euler angles over a serial port, or acting as a step-counting pedometer. The 9-DoF Razor’s MPU-9250 has a triple 3-axis sensors configuration: an accelerometer to measure linear acceleration, a gyroscope to measure angular rotation, and a magnetometer to measure magnetic field vectors around it. The parts can be placed on two Laser cut or 3D printed plates. One plate is dedicated to carrying the Nvidia Jetson TX2 dev kit, while the rest of the parts can be distributed between the second plate and the car chassis (see Figure 1(b) for components). The main computer runs ROS on top of the Ubuntu operating system. ROS environment serves the purpose of making our software stack modular, i.e. separate software modules can be developed for different tasks, such as the computer vision system, the autonomous driving system, or the motors control system. Educational approach Within the 2018 summer internship program, a group of 15 undergraduate computer engineering students was selected to be taught complex robot tasks on the racecar mini autonomous developer kit. The applied reproducible educational program includes theoretical lectures and laboratory assignments that allow students to practice their theoretical knowledge and get hands-on experience. A literate programming approach (via Jupyter notebooks) with the Python programming language is used for programming the vehicle. Originally introduced in 2011 as IPython notebooks, Jupyter notebooks37 is an open-source interactive web-based application streamlining and simplifying literate programming.38,39 Jupyter notebooks supported only Python at first, as the early 346 International Journal of Electrical Engineering & Education 57(4) Figure 2. Thresholding example of Jupyter Notebook from assignment labs. name implies, then extended the support to a large number of programming languages, by adding the appropriate Jupyter kernels. The kernels communicate with the system using Jupyter’s communication protocol, the so-called Jupyter messaging protocol. A notebook is a JavaScript Object Notation (JSON) document, which is parsed and displayed in a human-friendly format (see Figures 2 and 3). Each laboratory assignment notebook is a document containing the description of an image-analysis workflow using equations, text, the executable implementation of the workflow, and figures and tables generated by running the code. The code is split into logical sections called “code cells” or “code chunks” that can be executed interactively, providing a natural setting for modifying the code and experimenting with parameter values and seeing their effect on algorithm performance. Another benefit of running individual code chunks is that one can choose to run some parts only once, while others can be repeatedly executed. For example, while a dataloading cell only needs to run once, an algorithmic cell could run several times to compare different approaches to solving the same problem before settling on the best one. Eken et al. 347 Figure 3. Segmentation example of Jupyter Notebook from assignment labs. Theoretical lectures This subsection introduces the seven theoretical lectures. Each lecture includes a brief description of the learning objectives. Table 2 shows the recommended road map for instructors, with each assignment and its appropriate lecture. LEC0—ROS fundamentals. This lecture covers the first basics of commanding the vehicle actuators (steering and motor). The students get familiar with how to obtain data from the Lidar, the IMU, and the wheel odometer (through the VESC). They also learn how to work with ROS40 and its components such as nodes, topics, messages, services, etc. ROS is a Berkeley Software Distribution (BSD)-licensed software for controlling robotic components on a PC. An ROS consists of a number of independent nodes that communicate with each other using a publish/subscribe messaging pattern.41 For example, a node might implement a particular sensor’s driver, and publishes its acquired data in a stream of 348 International Journal of Electrical Engineering & Education 57(4) Figure 4. Examples from the data set. (a) Sample 1, (b) sample 2, (c) sample 3, (d) sample 4, (e) sample 5, and (f) sample 6. messages. These messages are then available for consumption by all other nodes that subscribe to that stream, including filters, loggers, and also higher level systems such as pathfinding, collision detection, etc. Our software modules running on the Jetson42 are structured as ROS nodes. The nodes for the LIDAR and other sensors are part of the software package given to students. Students integrate existing software modules (drivers for reading sensor data, and their custom algorithms) to quickly develop a complete autonomous system. LEC1—Control systems theory. Many engineered systems have an input affecting the value of some output. If we are interested in regulating the values of the output, as we frequently are, we become interested in algorithms to adjust the inputs accordingly. This will be facilitated by making feedback about the current state of the output available to the control algorithm, though it is also possible to try to “blindly” adjust the input. Regulation algorithms and their implementations are called control systems. The difficulty of the control problem depends on certain factors including the physical complexity of the system, its time dynamics and those of the desired output trajectory, the resolution and latency of the available Eken et al. 349 Figure 5. An example from dataset and its XML file. (a) A sample of parking traffic sign (b) and its annotated XML file. Table 2. Syllabus for teaching mini autonomous race car programming. Week Lecture Laboratory assignment(s) 1 1 2,3 LEC0—ROS fundamentals LEC1—Control Systems Theory LEC2—Introduction to Computer Vision and OpenCV 2,3 LEC3—Deep Learning Fundamentals and Tools LEC4—Object Detection and Classification LEC5—Visual Servoing LEC6—Localization Testing and deployment of all assignments on the mini autonomous vehicle LAB0—ROS Development with Python LAB1—Bang-Bang and Safety Controllers LAB2—OpenCV Basics LAB3—Streaming Video with OpenCV LAB4—Semantic Segmentation LAB5—Blob Detection using the ZED Camera LAB6—End-to-End Learning for Steering Autonomous Vehicle LAB7—Traffic Sign Classification and Localization 2,3 4 5 6,7,8 LAB8—Parking in front of a Visual Marker LAB9—Histogram Filters and 2D Kalman Filter feedback sensors, etc.43 We study a collection of common and relatively straightforward control techniques such as feed-forward control, Bang–Bang control, Proportional–Integral–Derivative (PID) control, differential drive velocity control.44 350 International Journal of Electrical Engineering & Education 57(4) LEC2—Introduction to computer vision and OpenCV. An important part of the human driving experience is keeping an eye on the environment, as the safety of the driver and that of many other people depends on that. We specifically try to spot any obstacles such as pedestrians and other cars and evade them. Similarly, autonomous vehicles must leverage their sensors and intelligence to perform the same task of detecting obstacles as it reinforces the car’s understanding of its environment. Other vehicles on the road represent one of the most important obstacles to deal with, as they are most likely the biggest and the most frequently encountered objects in our lane or neighboring ones. Within this lecture, students start by learning basic operations on image processing with OpenCV45 such as accessing pixel values and modifying them, accessing image properties, setting region of Image (ROI), splitting and merging images, blurring, finding contours, line, and shape detection. Then they go on learning more advanced concepts applicable to autonomous driving. LEC3—Deep learning fundamentals and tools. In 2005, Sebastian Thrum and his team won the DARPA Grand Challenge46 using machine learning, which ignited a huge interest in machine learning and deep learning for developing autonomous vehicles. Today, continuous advances in deep learning technology keep pushing the autonomous vehicle industry forward. Within this lecture, students learn how to train a deep neural network to correctly classify newly introduced images, deploy their models into applications, find ways to improve the performance of these applications, deploy end-to-end deep learning models for autonomous driving. They also get familiar with frameworks like Tensorflow,47 the most famous library used in production for deep learning models and Keras48 which is a highlevel API built on TensorFlow. LEC4—Object detection and classification. Perception, which is the task of environment sensing including object classification, detection, 3D position estimation, and simultaneous localization and mapping (SLAM),49 is as important to autonomous vehicles as it is to human drivers. Without it, the vehicle would not be able to navigate through any kind of obstacles. And that brings us to the subtask of object detection, for which we need accurate and reliable algorithms to deploy on the autonomous vehicle. Within this lecture, students learn OverFeat,50 VGG16,51 and YOLO52 models. LEC5—Visual servoing. Visual servoing is the use of cameras and a computer vision system to control the motion of a robot. Students leverage what they learned about OpenCV and ROS so far combined with image-based visual servoing to program the racecar to park in front of solid-color objects, like orange cones or specific traffic signs. LEC6—Localization. Navigation is a key and challenging task for mobile robots. To perform navigation successfully, four subtasks are required: perception, Eken et al. 351 localization, cognition, and motion control. Localization is the robot’s ability to determine its position relative to the environment, and it is the area that has received the most research attention in the past decade, allowing for significant advances to be made. The use of localization techniques enables a robot to find two basic pieces of information: what are the robot’s current location and orientation within its environment? This lecture also covers subjects ranging from histogram filters to N-dimensional Kalman Filters as applied in localization. Laboratory assignments This section introduces the 10 laboratory experiments53 accompanying the lectures. LAB0—ROS development with Python. Typically, the entire software functionality of a robot is divided using ROS into different modules, each running over one or multiple nodes. This laboratory assignment covers how to write a simple ROS application consisting of a publisher and subscriber node in python and getting familiar with concepts like topics and messages. Node is the ROS term for an executable that is connected to the ROS network. Topics are named buses over which nodes exchange messages. A message is a simple data structure, comprising typed fields such as Bool, Int32, String, etc. In this laboratory, students will implement a publisher ROS node that generates a random number between 0 and 10 and write a subscriber ROS node that will square the generated number and display the result. LAB1—Bang-bang and safety controllers. The LIDAR sensor provides a cloud of distance measurements in a polar form to the system. Using this data, a system can map the environment around it and navigate it accordingly. In this lab, students are asked to implement a Bang-Bang safety controller as an “emergency stop” node using ROS and Python. The node will receive the LIDAR data through the LIDAR sensor topic. If the distances measurements right in front of the car are below a specified threshold, the node will recognize that there is an obstacle and publish a stop command to the SAFETY topic. The car, in turn, will stop as soon as it receives that command preventing the collision. LAB2—OpenCV basics. In this lab students learn how to use OpenCV for basic operations such as reading an image from a file, displaying an image on the screen, gray-scale conversion, mirroring, resizing, cropping, image segmentation, and shuffling. Also, they use the MatPlotLib library54 to display images. LAB3—Streaming video with OpenCV. In this lab, students learn how to send a continuous stream of images with ROS as a topic. Our goal here is learning how to distribute sensor data from one node to several other nodes that need the data. 352 International Journal of Electrical Engineering & Education 57(4) As is the case with the camera in our project, and the several modules that will be using its pictures to perform different tasks. LAB4—Semantic segmentation. Image semantic segmentation is the partitioning of an image into semantically meaningful areas or classes. In a sense, it is similar to image classification but applied at a pixel level. Semantic segmentation has been used in a wide range of applications ranging from robotics to augmented reality, including in autonomous driving. For the latter application, a scene could be segmented into streets, traffic signs, street markings, cars, pedestrians, or sidewalks, all of which are important data for an autonomous vehicle. In this laboratory, students will train a SegNet55 on the CamVid data set56 for road scenes, and then try it on the data from the car’s camera. LAB5—Blob detection using the ZED camera. The goal of this lab is to learn how to use the stereo camera with the onboard GPU to influence the vehicle’s control. This is possible thanks to computer vision algorithms implemented in the OpenCV library which gives a robot a sense of perception in its field of view. Using this perception to locate targets, called “blobs,” and track them, the students are asked to follow and ultimately make decisions based on the position of these blobs. The task of blob detection can be summarized with the following computer vision operations: • The first step is converting the frames from RGB (red, green, blue) into HSV (hue, saturation, value) color space. The differences can be understood from the names; while RGB stores each pixel as a combination of red, green, and blue values, HSV describes them as a combination of hue (angle around the color wheel), saturation (how “pure” the color is), and value (how vibrant the color is). • A threshold is applied to the image to filter the desired blob out with a certain color range. Filtering red for example is achieved by keeping only pixels with values between HSV (2.95%, 47%) and HSV (20, 100%, 100%). • The image is eroded to remove noise and smooth edges. • Outlines around the blobs are highlighted using The OpenCV findContours function. • Smaller contours usually result from noise, so we drop them and keep only the one with the largest area. • A polygon encapsulating the resulting contour is approximated using OpenCV’s approxPolyDP. If it fits the shape of the desired blob, we keep it, otherwise, we drop it. • The blob is described as a single position in terms of the center of the final polygon. LAB6—End-to-end learning for steering autonomous vehicle. Vehicle steering, as simple as it seems, is one of the most challenging aspects of an Artificial Intelligence (AI) robotic system. That is why researchers have focused on training deep neural Eken et al. 353 Table 3. Sample from data set. Image File name Speed Angle 20 118 151 225 313 /home/jetson/racecar/056/00020.jpg /home/jetson/racecar/026/00118.jpg /home/jetson/racecar/026/00151.jpg /home/jetson/racecar/026/00225.jpg /home/jetson/racecar/026/00341.jpg 0.111419 0.625495 0.550052 0.500523 0.458846 0.0 0.11 0.24 0.17 0.24 network models on visual data from front cameras paired with the steering angles.57 In this lab, students deploy different end-to-end deep learning models on the Racecar and evaluate their performance. Experimental analysis shows that different models successfully manage to steer the vehicle autonomously. The pipeline for this lab can be listed as follows: • Building data set: The vehicle’s speed and steering angle are controlled with a Logitech F710 joystick, an ROS module subscribes to the events of the joystick recording both these values as well as images taken from the ZED which are then stored in a directory as shown in Table 3. • Splitting data set into training and test samples. • Python generator is used to generate data for training on the fly, rather than storing everything in memory. Also, data augmentation techniques such as flipping, brightness change, and shifts are applied. • Deploying different end-to-end models after training and steering the vehicle based on data from the front view camera. LAB7—Traffic sign classification and localization. Recognizing traffic signs58 is another crucial task of autonomous driving. In this lab assignment, students are asked to achieve this using a real-time object detection and classification algorithm, namely Tiny-YOLO.59 Students implement an ROS node with their trained Tiny-YOLO model. The workflow of traffic signs detection can be broken up into two steps: (i) locating possible signs in a large image and (ii) classifying the cropped traffic signs if it is indeed relevant and recognized by the trained model or dismissing it if not. In this lab, students cover in detail the different steps needed to create a pipeline, which enables localization and classification of traffic signs. The pipeline itself looks as follows: • Building data set: same as in LAB6 (see Figure 4(a) to (f)). • Image annotation: we labeled the frames using an image annotation tool called LabelImg.60 The traffic signs are first encompassed in a rectangular bounding box then the resulting area was labeled according to the sign’s class. Annotations are stored in an XML file following the same format used by ImageNet and called PASCAL VOC. Each entry stores the coordinates, 354 • • • • • International Journal of Electrical Engineering & Education 57(4) width, height, and class label of its respective image (see Figure 5(a) and (b) for an example). Data normalization: Learning is more stable by reducing variability across the training data, the network will generalize better to novel data, and the inputs will fall within the useful range of your nonlinearities. Data augmentation: Image augmentation is a necessary step to improve the performance of the model and mitigate the downside of having a small training data set for our deep neural network. This is achievable by synthesizing training images using one or more processing methods such as random rotation, brightness change, and axis flips. imgaug61 is a powerful library implementing image augmentation techniques for machine learning. Getting Google Colab ready to use: Google Colaboratory, or shortly Colab,62 is a free cloud service offered by Google for machine learning education and research. It offers dedicated virtual machines equipped with NVIDIA Tesla K80 cards for GPU acceleration and comes with an online Jupyter notebook environment canceling the need for any setup on a local machine. Image classification using different pretrained models: In this part, students learn transfer learning and fine-tuning. Darknet-19 was used as an example pretrained model. Measuring network performance: Students put the model to the test with their testing dataset and evaluate various performance metrics like accuracy, loss, precision, etc. LAB8—Parking in front of a visual marker. The goal of this lab is binding the vehicle’s vision directly to its control. Allowing it to reliably navigate toward a visual marker. The visual marker is determined by blob detection which is covered in LAB5, what this lab adds on top of that is the ability to track the blob’s center along the horizontal axis and issue the appropriate Ackermann steering commands to direct the vehicle toward it. LAB9—Histogram filters and 2D Kalman filter. For this lab, students deal with filtering, which is a general method for estimating unknown variables from noisy observations over time. In particular, they learn the Bayes Filter and some of its variants: The Histogram Filter and the Kalman Filter. They can apply the algorithms to the robot localization problem. Evaluation of educational program and discussion of results To quantitatively assess the educational program’s impact on the above learning process and get a tangible idea on the students’ satisfaction level, an electronic survey was conducted at the last session of the summer internship program, in which students had to grade their satisfaction in 14 questions form according to the five-level Likert-type scale: strongly disagree, disagree, neutral, agree, and strongly agree. Thus, they were able to express their opinion about how useful Eken et al. 355 Table 4. Students’ survey questions. No. Q1 Q2 Q3 Q4 Q5 Q6 Q7 Q8 Q9 Q10 Q11 Q12 Q13 Q14 Question Learning value LAB0 enhanced my ability to understand the theoretical ROS lecture in a new way. LAB1 enhanced my ability to understand the theoretical feedback control systems lecture in a new way. LABs 2-5 enhanced my ability to understand the theoretical image processing lecture in a new way. LAB6 and LAB7 enhanced my ability to understand the theoretical deep learning lecture in a new way. LAB8 enhanced my ability to understand the theoretical visual servoing lecture in a new way. LAB9 enhanced my ability to understand the theoretical localization lecture in a new way. Value added The reproducible educational plan helped me to integrate different technologies into working systems solutions. I can learn similar topics in this way without such a training plan. I can prepare such a reproducible educational study for other lessons. Design and other issues The ideas and concepts within the reproducible educational plan were presented clearly and easy to follow. The time allocated to the course was enough. I was interested in this topic before I started to study it. I am interested in this topic after having studied it. I recommend this educational program to a fellow student. the program was in reinforcing and improving their knowledge of the field. This feedback from students is important because it gives instructors a good idea on which characteristics of their teaching approach should be modified or not. Allowing each new session to be an improvement over the previous one. Moreover, the interaction between students and educators ensures the effectiveness of the course, the maintenance of the quality of university education and permits them to reflect upon their experiences among others. Questionnaire items are structured in three subscales: “Learning value,” “Value added,” and “Design and other issues.” Following is an explanation of the aim of each item: • Learning value encompasses the questions that try to probe students’ perceptions of how effectively the reproducible educational program is designed for the purpose of teaching mini autonomous race car programming. • Value added tries to assess the ability to learn and prepare similar systems and how much of this ability is thanks to this program. • Design and other issues focus on students’ judgment of the clarity of the course and their interest in the program. 356 International Journal of Electrical Engineering & Education 57(4) Figure 6. Student responses for the reproducible educational plan survey (number of students ¼ 15). (a) Learning value (Q1–Q6) (b) value added (Q7–Q9) and (c) design and other issues (Q10–Q14). The full questionnaire is presented in Table 4. Answers are collected anonymously and the survey data are analyzed using Excel. The first six items make up the Learning value subscale. The subsequent three items make up the value added subscale. While the last five items relate to the design and other issues subscale. Figure 6(a) to (c) represents the responses of the 15 students who took part in the survey, the high percentage of “strongly agree” and “agree” answers indicates the success of this reproducible educational program for mini autonomous car programming. About the learning value, 60% of students agree or strongly agree, showing an important positive attitude toward the improvement of their abilities in topics such as ROS, feedback control systems, image processing basics, and visual servoing. Forty-seven percent of students agree or strongly agree that they can integrate different technologies, but a relatively large amount of students are neutral toward preparing such a reproducible educational notebook. Most of the students answered positively regarding the educational design and other issues section. Conclusions A reproducible educational program specifically focused on the teaching of mini autonomous car programming has been presented in this paper. The proposed Eken et al. 357 training program includes different lectures and laboratory assignments. Data from students’ feedback indicate the success in achieving the learning objectives of the program. The student response was satisfactory, showing a high degree of interest in the proposed approach as well as satisfaction with it. We will increase the number of students in the upcoming years for this reproducible open-source robotic education program. We plan to add a vehicle2vehicle communication module for better handling of closed roads and congested traffic scenarios. The locations of traffic accidents and jams are expected to be broadcast by other vehicles allowing the receivers to avoid such roads and plan alternative routes. Declaration of Conflicting Interests The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article. Funding The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This work was supported by Kocaeli University Scientific Research and Development Support Program (BAP) in Turkey under project number 2017/067HD. We would also like to thank OBSS for their support and OpenZeka for their training under MARC program. ORCID iD Süleyman Eken https://orcid.org/0000-0001-9488-908X References 1. Collberg C and Proebsting TA. Repeatability in computer systems research. Commun ACM 2016; 59: 62–69. 2. Hothorn T and Leisch F. Case studies in reproducibility. Brief Bioinformat 2011; 12: 288–300. 3. Ruvolo P. Dude, where’s my robot?: A localization challenge for undergraduate robotics. Proc AAAI ConfArtifIntell 31: 4798–4802. 4. Bihl T, Jenkins T, Cox C, et al. From lab to internship and back again: learning autonomous systems through creating a research and development ecosystem. In: Proc AAAI Conf Artif Intell 33: 9635–9643. 5. Casser V, Pirk S, Mahjourian R, et al. Depth prediction without the sensors: leveraging structure for unsupervised learning from monocular videos. In: Proc AAAI Conf Artif Intell 33: 8001–8008. 6. Chi L and Mu Y. Deep steering: learning end-to-end driving model from spatial and temporal visual cues. arXiv preprint arXiv:170803798 2017. 7. Fridman L, Jenik B and Reimer B. Arguing machines: perception control system redundancy and edge case discovery in real-world autonomous driving. arXiv preprint arXiv:171004459 2017. 358 International Journal of Electrical Engineering & Education 57(4) 8. Snider JM, et al. Automatic steering methods for autonomous automobile path tracking. Pittsburgh, PA: Robotics Institute, 2009. 9. Bohg J, Ciocarlie M, Civera J, et al. Big data on robotics. Big Data 2016; 4: 195–196. 10. Workshop at ICRA, http://www.heronrobots.com/EuronGEMSig/gem-sig-events/ icra2017workshoprrr (accessed 10 February 2019). 11. Workshop at ICRA in Brisbane, http://www.heronrobots.com/EuronGEMSig/gem-sigevents/ICRA2018Workshop (accessed 10 February 2019). 12. IEEE RAS summer school on deep learning for robot vision, http://robotvision2019. amtc.cl/ (accessed 10 February 2019). 13. Code ocean, https://codeocean.com/ (accessed 10 February 2019). 14. Bonsignorio F. A new kind of article for reproducible research in intelligent robotics [from the field]. IEEE Robot Automat Mag 2017; 24: 178–182. 15. Dahnoun N. Teaching electronics to first-year non-electrical engineering students. Int J Electr Eng Educ 2017; 54: 178–186. 16. Nourbakhsh I, Crowley K, Wilkinson K, et al. The educational impact of the robotic autonomy mobile robotics course. Technical Report CMU-RI-TR-03-29. Pittsburgh, PA: Carnegie Mellon University, 2003. 17. Nourbakhsh IR, Crowley K, Bhave A, et al. The robotic autonomy mobile robotics course: robot design, curriculum design and educational assessment. Auton Robots 2005; 18: 103–127. 18. de Gabriel JMG, Mandow A, Fernandez-Lozano J, et al. Using lego nxt mobile robots with labview for undergraduate courses on mechatronics. IEEE Trans Educ 2011; 54: 41–47. 19. Calnon M, Gifford CM and Agah A. Robotics competitions in the classroom: enriching graduate-level education in computer science and engineering. Global J Eng Educ 2012; 14: 6–13. 20. Bousaba NA, Conrad CMHJM and Cecchi V. Keys to success in the IEEE hardware competition. In: Proceedings of the American Society for Engineering Education Annual Conference, pp.1–18. 21. Chris C. Learning with first LEGO league. In: McBride R and Searson M (eds) Proceedings of society for information technology & teacher education international conference 2013. New Orleans, LA: Association for the Advancement of Computing in Education (AACE), pp. 5118–5124. 22. Fike H, Barnhart P, Brevik CE, et al. Using a robotics competition to teach about and stimulate enthusiasm for Earth science and other STEM topics. In EGU General Assembly Conference Abstracts, EGU General Assembly Conference Abstracts 18: EPSC2016–10. 23. National robotics challenge, https://www.thenrc.org/ (accessed 10 February 2019). 24. Engineering and robotics learned young (early), http://www.earlyrobotics.org/ (accessed 10 February 2019). 25. Self-driving fundamentals: Featuring Apollo, https://www.udacity.com/course/self-driv ing-car-fundamentals-featuring-apollo–ud0419 (accessed 10 February 2019). 26. Mit 6.s094: Deep learning for self-driving cars, https://selfdrivingcars.mit.edu/ (accessed 10 February 2019). 27. McLurkin J, Rykowski J, John M, et al. Using multi-robot systems for engineering education: teaching and outreach with large numbers of an advanced, low-cost robot. IEEE Trans Educ 2013; 56: 24–33. Eken et al. 359 28. Danahy E, Wang E, Brockman J, et al. Lego-based robotics in higher education: 15 years of student creativity. Int J Adv Robot Syst 2014; 11: 27. 29. Afari E and Khine M. Robotics as an educational tool: Impact of Lego mindstorms. International Journal of Information and Education Technology 2017; 7: 437–442. 30. Crenshaw TL and Beyer S. Upbot: a testbed for cyber-physical systems. In: Proceedings of the 3rd international conference on cyber security experimentation and test. CSET’10. Berkeley, CA: USENIX Association, pp. 1–8. 31. Kelly J, Binney J, Pereira A, et al. Just add wheels: leveraging commodity laptop hardware for robotics and AI education. In: AAAI 2008 AI Education Colloquium. Chicago, IL. Also appears in AAAI Technical Report WS-08-02. Menlo Park, CA: AAAI Press, pp. 50–55. 32. Karaman S, Anders A, Boulet M, et al. Project-based, collaborative, algorithmic robotics for high school students: programming self-driving race cars at MIT. In: 2017 IEEE integrated STEM education conference (ISEC). pp. 195–203. IEEE. 33. VESC—open source esc, http://vedder.se/2015/01/vesc-open-source-esc/ (accessed 10 February 2019). 34. Stereo labs, https://www.stereolabs.com/zed/ (accessed 10 February 2019). 35. Scanse, https://github.com/scanse (accessed 10 February 2019). 36. Sparkfun, https://www.sparkfun.com/products/retired/10736 (accessed 10 February 2019). 37. Jupyter notebook, https://jupyter.org/ (accessed 10 February 2019). 38. Kluyver T, Ragan-Kelley B, Perez F, et al. Jupyter notebooks—a publishing format for reproducible computational workflows. In: Loizides F and Scmidt B (eds) Positioning and power in academic publishing: players, agents and agendas. Amsterdam: IOS Press, pp. 87–90. 39. Shen H. Interactive notebooks: sharing the code. Nature 2014; 515: 151–152. 40. Ros, http://www.ros.org/ (accessed 10 February 2019). 41. Kul S, Eken S and Sayar A. Distributed and collaborative real-time vehicle detection and classification over the video streams. Int J Adv Robot Syst 2017; 14: 1729881417720782. 42. Jetson tx2 module, https://developer.nvidia.com/embedded/buy/jetson-tx2 (accessed 10 February 2019). 43. Twigg P, Ponnapalli P and Fowler M. Workshop problem-solving for improved student engagement and increased learning in engineering control. Int J Electr Eng Educ 2018; 55: 120–129. 44. Skruch P, et al. Control systems in semi and fully automated cars. In Mitkowski W, Kacprzyk J and Oprzedkiewicz K (eds) Trends in advanced intelligent control, optimization and automation. Cham: Springer International Publishing, pp. 155–167. ISBN 978-3-319-60699-6. 45. Opencv, https://opencv.org/ (accessed 10 February 2019). 46. Darpa grand challenge, https://cs.stanford.edu/group/roadrunner/old/index.html (accessed 10 February 2019). 47. Tensorflow, https://github.com/tensorflow/tensorflow (accessed 10 February 2019). 48. Keras, https://github.com/keras-team/keras (accessed 10 February 2019). 49. Montemerlo M, Thrun S, Koller D, et al. Fastslam: a factored solution to the simultaneous localization and mapping problem. In: In Proceedings of the AAAI national conference on artificial intelligence. Menlo Park, CA: AAAI, pp. 593–598. 360 International Journal of Electrical Engineering & Education 57(4) 50. Sermanet P, Eigen D, Zhang X, et al. Overfeat: integrated recognition, localization and detection using convolutional networks, https://arxiv.org/abs/1312.6229 (2013). 51. Simonyan K and Zisserman A. Very deep convolutional networks for large-scale image recognition. CoRR 2014; abs/1409.1556. 52. Redmon J, Divvala S, Girshick R, et al. You only look once: unified, real-time object detection. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp. 779–788. 53. Racecar-training, https://github.com/husmen/racecar-training (accessed 10 February 2019). 54. Matplotlib, https://matplotlib.org/ (accessed 10 February 2019). 55. Segnet, http://mi.eng.cam.ac.uk/projects/segnet/ (accessed 10 February 2019). 56. Motion-based segmentation and recognition dataset, http://mi.eng.cam.ac.uk/research/ projects/VideoRec/CamVid/ (accessed 10 February 2019). 57. Karsli M, Satilmiş Y, Şara M, et al. End-to-end learning model design for steering autonomous vehicle. In: 2018 26th Signal processing and communications applications conference (SIU), pp. 1–4. IEEE. 58. Sat Ilm Iş Y, Tufan F, Şara M, et al. Cnn based traffic sign recognition for mini autonomous vehicles. In: Swiatek J, Borzemski L and Wilimowska Z (eds) Information systems architecture and technology: proceedings of 39th international conference on information systems architecture and technology—ISAT 2018. Cham: Springer International Publishing, pp. 85–94. ISBN 978-3-319-99996-8. 59. Redmon J and Farhadi A. Yolo9000: better, faster, stronger. In: 2017 IEEE conference on computer vision and pattern recognition (CVPR), pp. 7263–7271. 60. labelimg, https://github.com/tzutalin/labelImg (accessed 10 February 2019). 61. imgaug, https://github.com/aleju/imgaug (accessed 10 February 2019). 62. Colab, https://colab.research.google.com (accessed 10 February 2019).