To recap, our goal in this chapter is to create an application that will recognize what it sees. We will start by first capturing video frames, prepare these frames for our model, and finally feed them into a Core ML model to perform inference. Let's get started.
If you haven't done it already, download the latest code from the accompanying repository: https://github.com/packtpublishing/machine-learning-with-core-ml. Once downloaded, navigate to the directory Chapter3/Start/ObjectRecognition/
and open the project ObjectRecognition.xcodeproj
. Once loaded, you will see the skeleton project for this chapter, as shown in the following screenshot:
To help you navigate around the project, here is a list of core files/classes and their main functions:
VideoCapture
will be responsible for the management and handling of the camera, including capturing video framesCaptureVideoPreviewView.swift
contains the classCapturePreviewView
, which will be used...