Match Game Event Intro
This page gives an overview of the steps required during our live Microsoft Match Game Events.
Our short welcome video will introduce you to the event and how to get started
The following short list of items summarises the next actions you need to take to start developing your game playing bot and to enter your bot into the event's tournament:
- 4.After you have made some progress in developing your code, make sure it is submitted to the event's tournament. Having your code run in the tournament is the only way that we measure your participation in the event.
You should now have everything you need to get started. Watch the other videos below for more specific information.
There are more videos to show you how to compete in the event and how to modify the Template code.
Learn exactly where you need to add your code to develop your game playing bot. What information you will receive, and what information you need to return in order to play the game.
Entering your code into a tournament lets you find out how good your game playing bot is. Each event will have at least one tournament and the video below gives you a quick overview of how to make sure your code is registered to play.
The Online Code Editor is where you spend most of your time as you write the code for your automated game playing bot. It's also where you run the code to play the games. Find out all about the Online Code Editor and how to play games in this video
We give you template code that can already analyse images and identify which animal is in them using the Computer Vision API. The first step to playing the Match Game better is to update the code to identify and match landmarks in images.
The template code has comments to guide you in achieving this, but this video provides step by step instructions. If you follow the steps in the video, your code should be able to recognise both animals and landmarks and play a vastly improved game.
There are three types of image included in the tiles used in the Match Game - animals, landmarks and words. Once your code is successfully recognising animals and landmarks, you can use the additional Cognitive Services Computer Vision API, OCR (optical character recognition) to read the word on tiles which do not appear to have an animal or landmark present. This video explains how to implement OCR in your code and begin to recognise and match word type tiles. This will also give you the skills you need to implement the advanced strategy of reading the types of tiles from their backs and making your guesses more accurate as a result.
The Microsoft Cognitive Services Computer Vision Analyze API returns extensive information on the image that was analysed. This video shows how this information can be displayed within the Online Code Editor (OCE) and therefore the relevant components required to play the game can be extracted. Column three of the OCE is used to display the result returned and we explain how this can be interpreted into Python code in your solution to access the relevant parts of the data.
JSON objects are widely used to transfer data to and from API services. They are human readable text strings which adhere to a formal syntax which means they are also readable in software. JSON objects can be easily manipulated in Python code by converting them to or from dictionary objects. This video introduces the format of JSON objects, demonstrates how to convert them to and from Python dictionary objects, and gives examples of how to work with dictionaries in your code.
We recommend tackling the following steps in order as the best approach to improving your code:
- 1.Implement Landmark matching.
- 2.Implement text recognition and word matching
- 3.Read the tile backs to match tiles from the same category
- 4.Check if you can match tiles in the bonus category
- 5.Wait before matching any tiles to match tiles in consecutive Bonus Categories.