RoboCupSimData: A RoboCup soccer research dataset Olivia Michael1, Oliver Obst∗1, Falk Schmidsberger2, Frieder Stolzenburg2 1: Western Sydney University, Centre for Research in Mathematics 2: Harz University of Applied Sciences, Automation and Computer Sciences Department ∗: Corresponding author: o.obst@westernsydney.edu.au Abstract RoboCup is an international scientific robot competition in which teams of multiple robots compete against each other. Its different leagues provide many sources of robotics data, that can be used for further analysis and application of machine learning. This paper describes a large dataset from games of some of the top teams (from 2016 and 2017) in RoboCup Soccer Simulation League (2D), where teams of 11 robots (agents) compete against each other. Overall, we used 10 different teams to play each other, resulting in 45 unique pairings. For each pairing, we ran 25 matches (of 10 mins), leading to 1125 matches or more than 180 hours of game play. The generated CSV files are 17 GB of data (zipped), or 229 GB (unzipped). The dataset is unique in the sense that it contains both the ground truth data (global, complete, noise-free information of all objects on the field), as well as the noisy, local and incomplete percepts of each robot. These data are made available as CSV files, as well as in the original soccer simulator formats. 1 Introduction RoboCup is an international scientific robot competition in which teams of multiple robots compete against each other. The RoboCup soc- cer leagues provide platforms for a number of challenges in robotics research, including loco- motion, vision, real-time decision making, deal- ing with partial information, multi-robot co- ordination and teamwork. In RoboCup, sev- eral different leagues exist to emphasize spe- cific research problems by using different kinds of robots and rules. There are different soccer leagues in the RoboCup with different types and sizes of hardware and software: small size, mid- dle size, standard platform league, humanoid, 2D and 3D simulation (Kitano et al., 1997). In the soccer simulation leagues (Akiyama et al., 2015), the emphasis is on multi-robot team work with partial and noisy information, in real-time. Each of the robots is controlled by a separate program that receives sensor information from the simu- lator as an input, and, asynchronously to sensor input, also decides on a next action that is sent back to the simulator, several times a second. This complexity of the environment, with con- tinuous state and action spaces, together with the opportunity to compete against each other, makes RoboCup soccer an interesting testbed for multi-robot learning among many other applica- tions. To assist automated learning of team behav- ior, we provide a large dataset generated using 10 of the top participants in RoboCup 2016 or 2017. While it is possible to use the simulator for robot learning, we also generate additional data that is not normally available from playing other teams directly: We modified the simulator to record data from each robots local perspective, that is, with the restricted views that depend on each robots situation and actions, and also include the sensor noise. In addition, for every step of the game (100 ms), we recorded ground truth in- formation (such as positions and velocities) of 1 arXiv:1711.01703v1 [cs.AI] 6 Nov 2017 all objects on the soccer field, as well as basic actions of each robot. This ground truth infor- mation is usually recorded in a logfile, but not available to teams during a match. To create our dataset, we ran all pairings of the selected teams with 25 repetitions of each game, that is, 1125 games in total. With 11 robots in each team, a single game dataset consists of 22 local views plus a global (ground-truth) view. These views are made available as CSV files (comma-separated values). We also provide the original logfiles that include additional sen- sors, actions of each robot recorded as text files, and the software we created to patch the simu- lator and to convert recordings into CSV. The provided data are useful for various dif- ferent tasks including imitation learning (e.g., Ben Amor et al., 2013), learning or testing of self-localization (e.g., Olson, 2000), predictive modeling of behavior, transfer learning and re- inforcement learning (e.g., Taylor and Stone, 2009), and representation learning for time series data (Michael et al., 2018). The next sections de- scribe the environment, robots, and data in more detail. 2 Description of the environment The RoboCup Soccer Simulation Server rcssserver (Noda et al., 1998) is the software used for the annual RoboCup competitions that have been held since 1997. It is hosted at github.com/rcsoccersim/. We used the rcssserver version 15.3.0 to create the data for this paper. The simulator implements the (2D) physics and rules of the game, and also handles the interface to the programs controlling each player. By default, players use a 90 degree field of view, and receive visual information every 150 ms. Throughout the game, this frequency can actively be changed by each player individ- ually to 75 ms, 150 ms, or 300 ms, by changing the field of view to 45 degrees, 90 degrees, or 180 degrees, respectively. Visual information is transmitted in the form of (Lisp-like) lists of identified objects, with the level of detail of information depending on object distances. Potential objects include all other players on the field, the ball, and landmarks like goal posts, flags, and side lines. Each player also receives additional status information, including energy levels, referee decisions, and the state of the game, every 100 ms. Each robot can issue parameterized actions every 100 ms, to control its locomotion, the direction of its view, and its field of view. A more detailed description of the information transmitted can be found in the simulator manual (Chen et al., 2003). 3 Overview on the provided data In robotics, data collections often comprise li- dar data recorded by laser scans. This is very useful in many applications, e.g., field robotics, simultaneous localization and mapping (SLAM) (Tong et al., 2013), or specific purposes such as agriculture (Chebrolu et al., 2017). However, in many contexts, there is not only one but several robots which may be observed. The data from RoboCup that we consider here includes infor- mation about other robots in the environment and hence about the whole multi-robot system. Data from multi-agent systems like the RoboCup or the multi-player real-time strategy video game StarCraft (Lin et al., 2017) provide information on (simulated) environments as in robotics. However, in addition, they contain data on other agents and thus lay the basis for machine learning research to analyze and pre- dict agent behavior and strategies, which is im- portant in applications such as service robotics and multi-robot systems in general. To provide a diverse dataset, we include several teams from the last two RoboCup competitions, allowing for different behaviors and strategies. Perception and behavior of each robot dur- ing a game depends on the behavior of all other robots on the field. Game logfiles, that is, files containing ground truth information obtained from recording games, can be produced from the simulator, and are recorded in a binary format. Access to individual player percepts, however, is only possible from within the player code. To learn from behavior of other teams, it is useful to use the exact information that individual play- ers receive, rather than the global (and noise- free) information in recorded logfiles. We there- fore modified the simulator to additionally also record all local and noisy information as received by the robots on the field, in individual files for 2 2016 2017 CSU Yunlu CYRUS Gliders Fra-UNIted HELIOS HELIOS Ri-one HfutEngine MT Oxsy Figure 1: Teams from the RoboCup 2D soc- cer simulation world championships 2016 in Leipzig (Germany) and 2017 in Nagoya (Japan) selected for the dataset. each player. This information is stored in the same format as it is sent to players. We also provide code to translate these individual logs into CSV files that contain relative positions and velocities. We chose the top-ten teams from the RoboCup 2D soccer simulation world championships 2016 in Leipzig (Germany) and 2017 in Nagoya (Japan) (see Figure 1). Team binaries includ- ing further descriptions can be downloaded from: archive.robocup.info/Soccer/Simulation/2D. We played each team against each other team for 25 times, resulting in about 17 GB of data (zipped CSV files). For each game, we record the original logfiles including message logs. We also generate files with ground truth data as well as local player data in human-readable format. Fi- nally, we made our generating scripts available, so that they can be used to reproduce our results or to produce additional datasets using other robotic soccer teams. There is also a smaller subset of 10 games where the top-five teams play against each other once (163 MB CSV files plus 217 MB original logfiles). Our data is available at bitbucket.org/oliverobst/robocupsimdata/. 4 Description of the ground truth data According to rules of the world soccer association FIFA, a soccer pitch has the size of 105 m×68 m. This is adopted for the RoboCup soccer simula- tion league. Nevertheless, the physical boundary of the area that may be sensed by the robots has an overall size of 120 m × 80 m. For the localiza- tion of the robot players, the pitch is filled with several landmark objects as depicted in Fig. 2: flags (f), which are punctual objects, lines (l), the goal (g), and the penalty area (p). The origin of the coordinate system is the center point (c). The pitch is divided horizontally (x-direction) into a left (l) and right (r) half and vertically (y- direction) into a top (t) and a bottom (b) half. Additional numbers (10, 20, 30, 40, or 50) indi- cate the distance to the origin in meters. Since every soccer game takes place on the same pitch, there is only one file with infomation about the landmarks for all games that lists all these coor- dinates, given as a table in CSV format, with name landmarks.csv. For example, the row f r t,52.5,34 says that the right top flag of the pitch has the (x, y)-coordinates (52.5, 34). Further table files provide information about the respective game. The names of all these files – all naming conventions are summarized in Fig. 3 – contain the names of the competing teams, the final scores for each team, possibly extended by the result of a penalty shootout, a time stamp (when the game was recorded), and some further identifier. The central SoccerServer (Chen et al., 2003) controls every virtual game with built-in physics rules. When a game starts, the server may be configured by several parameters which are col- lected in one file with the identifier parameters. For example, the row ball_decay,0.94 denotes that the ball speed decreases by the specified fac- tor (cf. Stolzenburg et al., 2002). However, from a robotics point of view, most of the information in this file is not very relevant, like the stamina of the robots, the noise model, or the format of the coach instructions. We thus skip further details here. A soccer simulation game in the RoboCup 2D simulation league lasts 10 mins in total, divided into 6000 cycles where the length of each cycle is 100 ms. Logfiles comprise information about the game, in particular about the current posi- tions of all players and the ball including their velocity and orientation for each cycle. This in- formation is collected for the whole game in a table with the identifier groundtruth. For each time point, the play mode (e.g. kickoff), the cur- rent ball position coordinates and its velocity is listed. Furthermore, the positions and velocities 3 Figure 2: Flags and lines in the robotic soccer simulation (cf. Chen et al., 2003). 1. landmarks: landmarks.csv (static information for all games, 1 file) 2. game data: