Skip to content

Reference

Available Models

See available detection models in open-image-models and OCR models in fast-plate-ocr.

FastALPR

FastALPR package.

ALPR

Automatic License Plate Recognition (ALPR) system class.

This class combines a detector and an OCR model to recognize license plates in images.

Source code in fast_alpr/alpr.py
class ALPR:
    """
    Automatic License Plate Recognition (ALPR) system class.

    This class combines a detector and an OCR model to recognize license plates in images.
    """

    def __init__(
        self,
        detector: BaseDetector | None = None,
        ocr: BaseOCR | None = None,
        detector_model: PlateDetectorModel = "yolo-v9-t-384-license-plate-end2end",
        detector_conf_thresh: float = 0.4,
        detector_providers: Sequence[str | tuple[str, dict]] | None = None,
        detector_sess_options: ort.SessionOptions = None,
        ocr_model: OcrModel | None = "global-plates-mobile-vit-v2-model",
        ocr_device: Literal["cuda", "cpu", "auto"] = "auto",
        ocr_providers: Sequence[str | tuple[str, dict]] | None = None,
        ocr_sess_options: ort.SessionOptions | None = None,
        ocr_model_path: str | os.PathLike | None = None,
        ocr_config_path: str | os.PathLike | None = None,
        ocr_force_download: bool = False,
    ) -> None:
        """
        Initialize the ALPR system.

        Parameters:
            detector: An instance of BaseDetector. If None, the DefaultDetector is used.
            ocr: An instance of BaseOCR. If None, the DefaultOCR is used.
            detector_model: The name of the detector model or a PlateDetectorModel enum instance.
                Defaults to "yolo-v9-t-384-license-plate-end2end".
            detector_conf_thresh: Confidence threshold for the detector.
            detector_providers: Execution providers for the detector.
            detector_sess_options: Session options for the detector.
            ocr_model: The name of the OCR model from the model hub. This can be none and
                `ocr_model_path` and `ocr_config_path` parameters are expected to pass them to
                `fast-plate-ocr` library.
            ocr_device: The device to run the OCR model on ("cuda", "cpu", or "auto").
            ocr_providers: Execution providers for the OCR. If None, the default providers are used.
            ocr_sess_options: Session options for the OCR. If None, default session options are
                used.
            ocr_model_path: Custom model path for the OCR. If None, the model is downloaded from the
                hub or cache.
            ocr_config_path: Custom config path for the OCR. If None, the default configuration is
                used.
            ocr_force_download: Whether to force download the OCR model.
        """
        # Initialize the detector
        self.detector = detector or DefaultDetector(
            model_name=detector_model,
            conf_thresh=detector_conf_thresh,
            providers=detector_providers,
            sess_options=detector_sess_options,
        )

        # Initialize the OCR
        self.ocr = ocr or DefaultOCR(
            hub_ocr_model=ocr_model,
            device=ocr_device,
            providers=ocr_providers,
            sess_options=ocr_sess_options,
            model_path=ocr_model_path,
            config_path=ocr_config_path,
            force_download=ocr_force_download,
        )

    def predict(self, frame: np.ndarray | str) -> list[ALPRResult]:
        """
        Returns all recognized license plates from a frame.

        Parameters:
            frame: Unprocessed frame (Colors in order: BGR) or image path.

        Returns:
            A list of ALPRResult objects containing detection and OCR results.
        """
        if isinstance(frame, str):
            img_path = frame
            frame = cv2.imread(img_path)
            if frame is None:
                raise ValueError(f"Failed to load image from path: {img_path}")

        plate_detections = self.detector.predict(frame)
        alpr_results = []
        for detection in plate_detections:
            bbox = detection.bounding_box
            x1, y1 = max(bbox.x1, 0), max(bbox.y1, 0)
            x2, y2 = min(bbox.x2, frame.shape[1]), min(bbox.y2, frame.shape[0])
            cropped_plate = frame[y1:y2, x1:x2]
            ocr_result = self.ocr.predict(cropped_plate)
            alpr_result = ALPRResult(detection=detection, ocr=ocr_result)
            alpr_results.append(alpr_result)
        return alpr_results

    def draw_predictions(self, frame: np.ndarray | str) -> np.ndarray:
        """
        Draws detections and OCR results on the frame.

        Parameters:
            frame: The original frame or image path.

        Returns:
            The frame with detections and OCR results drawn.
        """
        # If frame is a string, assume it's an image path and load it
        if isinstance(frame, str):
            img_path = frame
            frame = cv2.imread(img_path)
            if frame is None:
                raise ValueError(f"Failed to load image from path: {img_path}")

        # Get ALPR results
        alpr_results = self.predict(frame)

        for result in alpr_results:
            detection = result.detection
            ocr_result = result.ocr
            bbox = detection.bounding_box
            x1, y1, x2, y2 = bbox.x1, bbox.y1, bbox.x2, bbox.y2
            # Draw the bounding box
            cv2.rectangle(frame, (x1, y1), (x2, y2), (36, 255, 12), 2)
            if ocr_result is None or not ocr_result.text or not ocr_result.confidence:
                continue
            # Remove padding symbols if any
            plate_text = ocr_result.text
            confidence: float = (
                statistics.mean(ocr_result.confidence)
                if isinstance(ocr_result.confidence, list)
                else ocr_result.confidence
            )
            display_text = f"{plate_text} {confidence * 100:.2f}%"
            font_scale = 1.25
            # Draw black background for better readability
            cv2.putText(
                img=frame,
                text=display_text,
                org=(x1, y1 - 10),
                fontFace=cv2.FONT_HERSHEY_SIMPLEX,
                fontScale=font_scale,
                color=(0, 0, 0),
                thickness=6,
                lineType=cv2.LINE_AA,
            )
            # Draw white text
            cv2.putText(
                img=frame,
                text=display_text,
                org=(x1, y1 - 10),
                fontFace=cv2.FONT_HERSHEY_SIMPLEX,
                fontScale=font_scale,
                color=(255, 255, 255),
                thickness=2,
                lineType=cv2.LINE_AA,
            )
        return frame

__init__(detector=None, ocr=None, detector_model='yolo-v9-t-384-license-plate-end2end', detector_conf_thresh=0.4, detector_providers=None, detector_sess_options=None, ocr_model='global-plates-mobile-vit-v2-model', ocr_device='auto', ocr_providers=None, ocr_sess_options=None, ocr_model_path=None, ocr_config_path=None, ocr_force_download=False)

Initialize the ALPR system.

Parameters:

Name Type Description Default
detector BaseDetector | None

An instance of BaseDetector. If None, the DefaultDetector is used.

None
ocr BaseOCR | None

An instance of BaseOCR. If None, the DefaultOCR is used.

None
detector_model PlateDetectorModel

The name of the detector model or a PlateDetectorModel enum instance. Defaults to "yolo-v9-t-384-license-plate-end2end".

'yolo-v9-t-384-license-plate-end2end'
detector_conf_thresh float

Confidence threshold for the detector.

0.4
detector_providers Sequence[str | tuple[str, dict]] | None

Execution providers for the detector.

None
detector_sess_options SessionOptions

Session options for the detector.

None
ocr_model OcrModel | None

The name of the OCR model from the model hub. This can be none and ocr_model_path and ocr_config_path parameters are expected to pass them to fast-plate-ocr library.

'global-plates-mobile-vit-v2-model'
ocr_device Literal['cuda', 'cpu', 'auto']

The device to run the OCR model on ("cuda", "cpu", or "auto").

'auto'
ocr_providers Sequence[str | tuple[str, dict]] | None

Execution providers for the OCR. If None, the default providers are used.

None
ocr_sess_options SessionOptions | None

Session options for the OCR. If None, default session options are used.

None
ocr_model_path str | PathLike | None

Custom model path for the OCR. If None, the model is downloaded from the hub or cache.

None
ocr_config_path str | PathLike | None

Custom config path for the OCR. If None, the default configuration is used.

None
ocr_force_download bool

Whether to force download the OCR model.

False
Source code in fast_alpr/alpr.py
def __init__(
    self,
    detector: BaseDetector | None = None,
    ocr: BaseOCR | None = None,
    detector_model: PlateDetectorModel = "yolo-v9-t-384-license-plate-end2end",
    detector_conf_thresh: float = 0.4,
    detector_providers: Sequence[str | tuple[str, dict]] | None = None,
    detector_sess_options: ort.SessionOptions = None,
    ocr_model: OcrModel | None = "global-plates-mobile-vit-v2-model",
    ocr_device: Literal["cuda", "cpu", "auto"] = "auto",
    ocr_providers: Sequence[str | tuple[str, dict]] | None = None,
    ocr_sess_options: ort.SessionOptions | None = None,
    ocr_model_path: str | os.PathLike | None = None,
    ocr_config_path: str | os.PathLike | None = None,
    ocr_force_download: bool = False,
) -> None:
    """
    Initialize the ALPR system.

    Parameters:
        detector: An instance of BaseDetector. If None, the DefaultDetector is used.
        ocr: An instance of BaseOCR. If None, the DefaultOCR is used.
        detector_model: The name of the detector model or a PlateDetectorModel enum instance.
            Defaults to "yolo-v9-t-384-license-plate-end2end".
        detector_conf_thresh: Confidence threshold for the detector.
        detector_providers: Execution providers for the detector.
        detector_sess_options: Session options for the detector.
        ocr_model: The name of the OCR model from the model hub. This can be none and
            `ocr_model_path` and `ocr_config_path` parameters are expected to pass them to
            `fast-plate-ocr` library.
        ocr_device: The device to run the OCR model on ("cuda", "cpu", or "auto").
        ocr_providers: Execution providers for the OCR. If None, the default providers are used.
        ocr_sess_options: Session options for the OCR. If None, default session options are
            used.
        ocr_model_path: Custom model path for the OCR. If None, the model is downloaded from the
            hub or cache.
        ocr_config_path: Custom config path for the OCR. If None, the default configuration is
            used.
        ocr_force_download: Whether to force download the OCR model.
    """
    # Initialize the detector
    self.detector = detector or DefaultDetector(
        model_name=detector_model,
        conf_thresh=detector_conf_thresh,
        providers=detector_providers,
        sess_options=detector_sess_options,
    )

    # Initialize the OCR
    self.ocr = ocr or DefaultOCR(
        hub_ocr_model=ocr_model,
        device=ocr_device,
        providers=ocr_providers,
        sess_options=ocr_sess_options,
        model_path=ocr_model_path,
        config_path=ocr_config_path,
        force_download=ocr_force_download,
    )

draw_predictions(frame)

Draws detections and OCR results on the frame.

Parameters:

Name Type Description Default
frame ndarray | str

The original frame or image path.

required

Returns:

Type Description
ndarray

The frame with detections and OCR results drawn.

Source code in fast_alpr/alpr.py
def draw_predictions(self, frame: np.ndarray | str) -> np.ndarray:
    """
    Draws detections and OCR results on the frame.

    Parameters:
        frame: The original frame or image path.

    Returns:
        The frame with detections and OCR results drawn.
    """
    # If frame is a string, assume it's an image path and load it
    if isinstance(frame, str):
        img_path = frame
        frame = cv2.imread(img_path)
        if frame is None:
            raise ValueError(f"Failed to load image from path: {img_path}")

    # Get ALPR results
    alpr_results = self.predict(frame)

    for result in alpr_results:
        detection = result.detection
        ocr_result = result.ocr
        bbox = detection.bounding_box
        x1, y1, x2, y2 = bbox.x1, bbox.y1, bbox.x2, bbox.y2
        # Draw the bounding box
        cv2.rectangle(frame, (x1, y1), (x2, y2), (36, 255, 12), 2)
        if ocr_result is None or not ocr_result.text or not ocr_result.confidence:
            continue
        # Remove padding symbols if any
        plate_text = ocr_result.text
        confidence: float = (
            statistics.mean(ocr_result.confidence)
            if isinstance(ocr_result.confidence, list)
            else ocr_result.confidence
        )
        display_text = f"{plate_text} {confidence * 100:.2f}%"
        font_scale = 1.25
        # Draw black background for better readability
        cv2.putText(
            img=frame,
            text=display_text,
            org=(x1, y1 - 10),
            fontFace=cv2.FONT_HERSHEY_SIMPLEX,
            fontScale=font_scale,
            color=(0, 0, 0),
            thickness=6,
            lineType=cv2.LINE_AA,
        )
        # Draw white text
        cv2.putText(
            img=frame,
            text=display_text,
            org=(x1, y1 - 10),
            fontFace=cv2.FONT_HERSHEY_SIMPLEX,
            fontScale=font_scale,
            color=(255, 255, 255),
            thickness=2,
            lineType=cv2.LINE_AA,
        )
    return frame

predict(frame)

Returns all recognized license plates from a frame.

Parameters:

Name Type Description Default
frame ndarray | str

Unprocessed frame (Colors in order: BGR) or image path.

required

Returns:

Type Description
list[ALPRResult]

A list of ALPRResult objects containing detection and OCR results.

Source code in fast_alpr/alpr.py
def predict(self, frame: np.ndarray | str) -> list[ALPRResult]:
    """
    Returns all recognized license plates from a frame.

    Parameters:
        frame: Unprocessed frame (Colors in order: BGR) or image path.

    Returns:
        A list of ALPRResult objects containing detection and OCR results.
    """
    if isinstance(frame, str):
        img_path = frame
        frame = cv2.imread(img_path)
        if frame is None:
            raise ValueError(f"Failed to load image from path: {img_path}")

    plate_detections = self.detector.predict(frame)
    alpr_results = []
    for detection in plate_detections:
        bbox = detection.bounding_box
        x1, y1 = max(bbox.x1, 0), max(bbox.y1, 0)
        x2, y2 = min(bbox.x2, frame.shape[1]), min(bbox.y2, frame.shape[0])
        cropped_plate = frame[y1:y2, x1:x2]
        ocr_result = self.ocr.predict(cropped_plate)
        alpr_result = ALPRResult(detection=detection, ocr=ocr_result)
        alpr_results.append(alpr_result)
    return alpr_results

ALPRResult dataclass

Dataclass to hold the results of detection and OCR for a license plate.

Source code in fast_alpr/alpr.py
@dataclass(frozen=True)
class ALPRResult:
    """
    Dataclass to hold the results of detection and OCR for a license plate.
    """

    detection: DetectionResult
    ocr: OcrResult | None

BaseDetector

Bases: ABC

Source code in fast_alpr/base.py
class BaseDetector(ABC):
    @abstractmethod
    def predict(self, frame: np.ndarray) -> list[DetectionResult]:
        """Perform detection on the input frame and return a list of detections."""

predict(frame) abstractmethod

Perform detection on the input frame and return a list of detections.

Source code in fast_alpr/base.py
@abstractmethod
def predict(self, frame: np.ndarray) -> list[DetectionResult]:
    """Perform detection on the input frame and return a list of detections."""

BaseOCR

Bases: ABC

Source code in fast_alpr/base.py
class BaseOCR(ABC):
    @abstractmethod
    def predict(self, cropped_plate: np.ndarray) -> OcrResult | None:
        """Perform OCR on the cropped plate image and return the recognized text and character
        probabilities."""

predict(cropped_plate) abstractmethod

Perform OCR on the cropped plate image and return the recognized text and character probabilities.

Source code in fast_alpr/base.py
@abstractmethod
def predict(self, cropped_plate: np.ndarray) -> OcrResult | None:
    """Perform OCR on the cropped plate image and return the recognized text and character
    probabilities."""