Skip to main content

AI-POWERED DOCS

What do you want to know?

Improving Your Model Over Time

Your inspection is deployed. Now how do you keep it performing well as conditions change, new defect types appear, or specs evolve?

The Library: Your continuous improvement tool

Every capture the camera takes gets saved in the Library, along with the AI's prediction and confidence score. This is your gold mine for improvement.

Library view: browse captures, filter by prediction and confidence

Find what the AI got wrong

  1. Go to the Library
  2. Browse recent captures
  3. Look for two things:
    • Misses: images where the AI's prediction is clearly wrong (checkbox in the top left of each thumbnail)
    • Low-confidence predictions: scroll down on any image to see confidence values. Low confidence = the AI was unsure
Mislabeled images can significantly hurt accuracy

One mislabeled image in a small dataset has an outsized impact. With only 5 training images, a single wrong label corrupts 20% of your data. Always double-check labels before retraining, especially when your dataset is small.

Retrain with targeted data

  1. Select the images the AI got wrong or was unsure about
  2. Click "Add to Active Recipe Train Set"
  3. Fix the labels if needed
  4. Click Retrain

Adding Library images to your training set

Focus on misses and low-confidence captures, not random new data. This is the most efficient way to improve.

Inspect saved images with the Metadata Viewer

If you need to see exactly what the AI predicted on a specific image, upload the saved JPEG to the Metadata Viewer. OV cameras embed prediction results, confidence scores, and camera settings directly into each captured image.

Classifier improvement workflow

  1. Review Library images → find errors and low-confidence predictions
  2. Add them to the training set
  3. Relabel if needed
  4. Retrain
  5. Use Haystack to explore your data visually and find mislabeled images at scale

Segmenter improvement workflow

Segmenters take longer to label (pixel-level annotation), so there's a shortcut:

  1. Import problem images into the segmentation recipe
  2. Click Generate Predictions so the model pre-labels the new images as best it can
  3. Fix the predictions rather than labeling from scratch (much faster)
  4. Retrain
Use Haystack to spot mislabels at scale

Haystack is a visual data exploration tool that lets you browse your entire training set and quickly spot images that look out of place. As your dataset grows beyond 50-100 images, manual review becomes impractical. Haystack clusters similar images together, making mislabeled outliers stand out visually.

Reducing false positives in segmentation

If your segmenter is picking up a lot of small random blobs that shouldn't be there, you have two powerful tools at your disposal: filtering small blobs out of the pass/fail logic, and training the model to better understand what "normal" looks like.

Filter out small blobs with a minimum size threshold

The fastest fix is to tell the system to ignore blobs below a certain pixel area. Small noise blobs are almost always tiny compared to real defects, so a size filter eliminates them without affecting your actual detection.

In Basic Mode (IO Logic): Set the minimum blob area threshold in the pass/fail settings. Any blob smaller than this pixel count gets ignored entirely. Start with a value like 50 pixels and adjust upward until the noise disappears but real defects still trigger.

In Advanced Mode (Node-RED): Filter blobs by blob.pixel_count in your function node. For example, to ignore any blob under 50 pixels:

const minBlobSize = 50;
const allBlobs = msg.payload.segmentation.blobs;
const realBlobs = allBlobs.filter(blob => blob.pixel_count >= minBlobSize);
const results = realBlobs.length < 1; // Pass if no real blobs remain
msg.payload = results;
return msg;

Adjust the minBlobSize value based on your application. Look at the pixel counts of your noise blobs vs. real defects in the debug output to find the right cutoff.

Add good images to your training set

This is the single most effective way to reduce false positive detections, and it's the one people most often overlook.

The problem: if you only trained with images that contain defects, the model has never learned what a clean part looks like. It starts seeing "defects" in normal surface texture, lighting gradients, and minor variations -- all the small random blobs you're fighting.

How to fix it:

  1. Go to the Library and find images of good parts -- clean surfaces with no defects
  2. Add these images to the training set (Add to Active Recipe Train Set)
  3. In the labeling view, select the ROIs but do not add any defect annotations. Just include them as-is with no paint marks
  4. Retrain the model

By showing the model examples of what "normal" looks like, you're teaching it to stop flagging ordinary surface variation as defects. This dramatically reduces noise.

Guidelines:

  • Start with 3-5 good images and retrain. Test the results
  • If noise persists, add more. Aim for at least 20-30 good images mixed with your defect images
  • This is especially important if your initial training set only contained images with defects on them
  • Make sure the good images cover the range of normal variation you see in production (different lighting, slight color shifts, surface textures)
This is the #1 fix for noisy segmenters

Nine times out of ten, when someone complains about random small blobs appearing everywhere, the root cause is that the training set has no good images in it. Add them and retrain -- the improvement is usually immediate and significant.

Other tips for reducing noise

  • Improve lighting consistency. Inconsistent lighting creates shadows and highlights that the model can mistake for defects. Make sure the built-in LEDs are configured properly and external light sources aren't causing glare or shifting shadows
  • Check your template alignment. If the aligner isn't working well, the ROI shifts between captures. The model sees different background areas each time, which increases noise. Verify alignment is stable
  • Use Accurate (production) training mode. Development mode is fast but much less precise. Always retrain in production mode before judging segmentation quality
  • Iterate in small steps. Add 3-5 good images, retrain, test. Repeat until results are clean. This is faster than dumping 50 images in at once and hoping for the best

The philosophy

  • Never stop iterating. The AI has high learning capacity, and it keeps improving with 50, 100, even 500+ images
  • Target failures specifically. Don't randomly add data. Add the cases where the AI struggles
  • Check for mislabels regularly. As your dataset grows, mislabels become harder to spot but still damage accuracy

Accelerate with GenAI tools

Three AI-powered tools at tools.overview.ai can dramatically speed up your improvement cycles:

  • Defect Studio -- Generate photorealistic synthetic defect images up to 10,000x faster than waiting for real defects
  • Integration Builder -- Build production-ready Node-RED flows from plain English descriptions
  • AI Expert Helper -- Get 24/7 expert guidance on any camera question
The GenAI toolkit advantage

Together, these three tools can reduce your deployment time from days to hours. Generate synthetic training data instead of waiting for real defects, build integrations by describing what you want, and get instant expert guidance without waiting for support tickets.

See the full GenAI Toolkit guide