Tag: the-invisible-factory

  • Deep Dive 04 — Weak APQP/PPAP Discipline

    TL;DR

    When APQP/PPAP is treated as paperwork instead of a risk management system, launches suffer. Generic PFMEAs, mismatched control plans, and skipped validations create instability and customer distrust.

    Executive Snapshot

    APQP and PPAP create a common language for launching products right the first time. When treated as checklists, they fail. Weak discipline leaves gaps in risk analysis, control plans, and validation. The result: late changes, unstable launches, and hidden risks that surface in the field.

    Why it happens

    • Launch pressure drives shortcuts to hit dates.
    • Docs filled to satisfy customers rather than to drive risk thinking.
    • Suppliers treat PPAP as a one‑time hurdle instead of a living system.
    • Management confuses “submitted” with “complete.”

    How it shows up

    • Control plans don’t match actual processes.
    • PFMEAs are generic, copy‑paste documents.
    • MSA and capability studies skipped or placeholder data used.
    • PPAPs pass on paper but fail on the shop floor; problems resurface post‑SOP.

    Consequences

    • Unstable launches and late design/process changes.
    • Early field failures; customer confidence erodes.
    • Escalations, extra audits, containment costs.
    • Factory stays noisy with chronic issues.

    The Fix

    • Restore APQP/PPAP as a risk management system, not a document package.
    • Link PFMEA failure modes to real process data.
    • Ensure control plans reflect how the process actually runs.
    • Run MSAs and capability studies before submission.
    • Cross‑functional readiness reviews; PPAP proves capability and readiness.

    Root Causes (6M+E)

    DimensionTypical Issues
    MeasurementSkipped MSAs; weak capability data
    MethodsGeneric PFMEAs; control plans misaligned with reality
    MachinesRun‑at‑rate skipped; unproven equipment
    MaterialsSupplier data unverified; raw material variation ignored
    ManpowerTeams treat APQP as paperwork; poor PFMEA/control plan skills
    EnvironmentLaunch sites unprepared; gaps found after SOP

    Diagnostics & Quick Checks

    • Do PFMEAs link to real process characteristics and data?
    • Does the control plan reflect what operators actually do?
    • Are MSAs and capability studies credible?
    • Was run‑at‑rate performed with production‑intent tools, parts, staffing?
    • Did the cross‑functional team confirm readiness to ship?

    Acceptance Criteria

    • PFMEAs identify real risks and link to control plans.
    • Control plans match actual processes and operator instructions.
    • MSAs and capability studies confirm readiness.
    • Run‑at‑rate validates volume capability.
    • Cross‑functional sign‑off without containment.

    30/60/90‑Day Playbook

    0–30 days

    • Audit recent APQP/PPAPs for PFMEA, control plan, MSA, and capability gaps.
    • Run training refreshers on APQP/PPAP intent.
    • Close obvious gaps in measurement and control plan accuracy.

    31–60 days

    • Revise PFMEAs with real failure data.
    • Align control plans with operator instructions and SPC.
    • Run‑at‑rate with production‑intent tools, staffing, environment.

    61–90 days

    • Institutionalize layered APQP/PPAP reviews.
    • Add APQP metrics to program dashboards.
    • Require revalidation after significant process, material, or supplier changes.

    Sustain & Scale

    • Tie APQP/PPAP to launch gates and supplier management.
    • Audit the shop floor, not just documents.
    • Capture lessons learned to strengthen future launches.
    • Treat APQP/PPAP as a living risk‑control system.

    When discipline is strong, launches stabilize, credibility grows, and the factory begins to play itself.

    📄 Download this Deep Dive as PDF

    References & Further Reading

  • Deep Dive 03 — Overreliance on 100% Inspection

    TL;DR

    Inspection feels safe but it isn’t. Even the best inspectors miss defects, and inspection adds cost without preventing failure. The only real safety net is building quality into the process itself.

    Executive Snapshot

    100% inspection feels safe, but it isn’t. Humans are poor at repetitive detection tasks — fatigue, bias, and attention gaps guarantee that defects slip through. Inspection doesn’t prevent defects; it only sorts them — at high cost. The invisible factory builds quality upstream so inspection becomes a verification step, not a wall.

    Why it happens

    • Prevention feels hard; “more eyes on parts” is the default.
    • Immature controls or missing capability data.
    • Cultural belief that inspection equals quality; doubling inspectors seems like doubling assurance.

    How it shows up

    • Large teams at end‑of‑line; multiple re‑inspections.
    • Disputes about what was “missed.”
    • Cosmetic metric gains while scrap/rework costs skyrocket.
    • Low morale as inspectors take the blame for systemic issues.

    Consequences

    • Human detection rarely exceeds 80–85% — escapes continue.
    • Costs rise; focus drifts from prevention to policing.
    • Customer trust erodes despite visible inspection “armies.”

    The Fix

    • Shift from detection to prevention.
    • Invest in process capability, error‑proofing, and SPC on CTQs.
    • Use inspection selectively; keep it as verification.

    Root Causes (6M+E)

    DimensionTypical Issues
    MeasurementInconsistent criteria; fatigue‑driven errors; subjective judgement
    MethodsInspection replaces root cause elimination; redundant checks
    MachinesProcess instability left uncontrolled; burden on inspectors
    MaterialsIncoming variation sorted instead of improved
    ManpowerInspectors blamed for escapes; training used as patch
    EnvironmentPoor lighting, ergonomics, distractions during inspection

    Diagnostics & Quick Checks

    • Seed known defects and measure detection rate.
    • Feed inspection data back into PFMEAs/control plans.
    • Watch for double/triple inspections — a sign of mistrust.
    • Rotate inspectors to reduce fatigue.

    Acceptance Criteria

    • Automate detection or error‑proof CTQs where possible.
    • Inspection as a verification layer only.
    • Success = fewer inspectors as capability rises.

    30/60/90‑Day Playbook

    0–30 days

    • Map where 100% inspection is used and why.
    • Measure detection rates via seeded defects.
    • Start root cause analysis on top 2–3 defect types caught by inspection.

    31–60 days

    • Mistake‑proof or automate detection for critical defects.
    • Stabilize upstream processes with SPC and capability studies.
    • Reduce redundant inspections; redeploy resources into prevention.

    61–90 days

    • Audit inspection cost vs cost of poor quality; show impact.
    • Institutionalize error‑proofing checks in control plans and PFMEAs.
    • Celebrate inspector reduction as process health, not risk.

    Sustain & Scale

    • Treat inspection as a temporary crutch.
    • Embed lessons into PFMEAs, control plans, and supplier requirements.
    • Audit so that processes — not inspectors — carry the burden of quality.

    📄 Download this Deep Dive as PDF

    References & Further Reading

  • Deep Dive 02 — Unstable Processes (No SPC or Control Charts)

    Executive Snapshot
    Without stability, a process cannot be predicted or improved. When SPC and control charts are missing, teams
    chase noise and miss signals. Problems are discovered late, usually through scrap, rework, or customer complaints.
    SPC is the early warning system of the invisible factory. It separates common cause from special cause, and it gives
    operators the confidence to act before defects pile up. Without it, firefighting replaces foresight.


    Why it happens
    SPC often gets sidelined for three reasons. First, it is seen as ‘statistical jargon’ that only quality engineers
    understand. Second, managers view it as overhead compared to lagging metrics like yield or PPM. Third,
    implementation is done poorly — wrong subgrouping, noisy data, or charts without reaction plans — so operators
    lose trust. The result: SPC is abandoned, and teams run blind.


    How it shows up
    Unstable processes reveal themselves through constant firefighting. SPC charts, if they exist, show frequent rule
    violations. Capability studies give contradictory results between runs. Operators don’t know whether to adjust or
    wait. Meetings are filled with debate over whether a shift was real or just ‘noise.’ Meanwhile, bad parts slip through
    undetected until customers raise complaints.


    Consequences
    When SPC is weak or absent, process capability is misunderstood. Engineers waste time chasing false signals.
    Operators lose confidence in the data. Quality costs rise as problems are detected late in inspection or, worse, by
    the customer. Management ends up relying on lagging indicators — a dangerous illusion of control.


    The fix
    The fix is to anchor SPC on critical-to-quality (CTQ) characteristics. Use proper subgrouping that reflects how the
    process produces. Train operators in the basic rules for detection — trends, runs, and points beyond limits. Most
    importantly, tie every chart to a reaction plan that defines what to do within 1–2 hours of a signal. SPC is not
    paperwork; it is the nervous system of a stable factory.


    Root causes (6M+E)
    Dimension Typical issues
    Measurement No SPC on CTQs; wrong subgrouping; noisy or unstable gauges.
    Methods No control plan; unclear reaction plans; over-adjusting instead of letting process run.
    Machines Unstable settings; drift; poor maintenance hides real variation.
    Materials Inconsistent supply quality; uncontrolled lot-to-lot variation.
    Manpower Operators not trained in SPC interpretation; charts ignored or filled in after the fact.
    Environment Temperature, humidity, or vibration introduce uncontrolled variation.


    Diagnostics & quick checks

    • Confirm SPC charts exist for all CTQs in the control plan.
    • Check if charts show stable periods with only common cause variation.
    • Look for documented reaction plans tied to specific rule violations.
    • Verify operators know what to do when a signal occurs.
    • Audit whether reaction happens within 1–2 hours of signal.
      Acceptance criteria
      Stable charts show no rule violations and only common cause variation. Out-of-control signals should trigger
      documented reaction within 1–2 hours. Capability studies should be run only on stable processes. Operators and
      engineers should interpret signals consistently with SPC rules.
      30/60/90-day playbook
      0–30 days
    • Identify CTQs and set up basic control charts.
    • Train engineers and operators in SPC rules of detection.
    • Pilot SPC on one high-risk process; link to a simple reaction plan.
      31–60 days
    • Expand SPC to other CTQs; standardize subgrouping methods.
    • Develop clear reaction plans and embed them in work instructions.
    • Audit first responses to signals; track reaction time.
      61–90 days
    • Integrate SPC metrics into management reviews.
    • Tie SPC signals into continuous improvement projects.
    • Reduce reliance on end-of-line inspection by using SPC as early warning.
      Sustain & scale
      Sustain SPC by embedding it into layered audits and control plans. Link recurring signals to structured problem
      solving. Review reaction effectiveness regularly and adjust plans as needed. When SPC is treated as a living
      system, not a chart-filling exercise, the invisible factory gains foresight — and firefighting declines.
      References & further reading
      [1] NIST/SEMATECH e-Handbook of Statistical Methods — Control Charts
      https://www.itl.nist.gov/div898/handbook/pmc/section3/pmc3.htm
      [2] AIAG SPC Manual (2nd Edition)
      https://www.aiag.org/quality/automotive-core-tools/spc
      [3] ASQ Quality Resources — Statistical Process Control
      https://asq.org/quality-resources/statistical-process-control
  • Deep Dive 01 — No MSA (Gage R&R) on Critical Gauges

    Executive Snapshot
    A process is only as trustworthy as the data behind it. When measurement systems aren’t qualified, teams make
    decisions on noise. Charts look unstable, capability studies mislead, and operators disagree about what’s in or out of
    spec. Measurement Systems Analysis (MSA) is the gatekeeper. Bias, linearity, stability, and Gage R&R; tell us if the
    numbers reflect reality. Without it, ‘data-driven decisions’ are built on sand.


    Why it happens
    MSA often gets skipped for three reasons. First, launch pressure puts throughput ahead of metrology readiness.
    Second, calibration is confused with capability — a gage can be ‘in date’ yet still unreliable. Third, vendor specs are
    taken at face value without checking in the actual production context. Each shortcut looks harmless. In practice, they
    feed chronic firefighting.


    How it shows up
    On the floor, weak measurement systems show up quickly. SPC charts are noisy with false alarms. Capability
    studies swing from good to bad between runs. Operators dispute pass/fail calls. Re-inspections give different
    results on the same part. These aren’t process problems. They’re measurement problems disguised as process
    instability.


    Consequences
    When data can’t be trusted, everything downstream erodes. Engineers waste hours chasing phantom shifts. Scrap
    and rework pile up because of false rejects. Quality leaders defend bad metrics to customers. The factory loses
    credibility — inside and outside.


    The fix
    The fix is straightforward: run proper MSAs on critical-to-quality (CTQ) gauges. For variables, use about 10 parts ×
    3 operators × 2–3 repeats. For attributes, run at least 50 randomized trials with repeats. Check %GRR, ndc, and
    agreement rates. Fix obvious issues — resolution, fixturing, training, environment — then rerun until capable. MSA
    isn’t paperwork. It’s the license to trust your data.


    Root causes (6M+E)
    Dimension Typical issues
    Measurement Unqualified gages; inadequate resolution; wear; bias/linearity not checked.
    Methods No work instruction; inconsistent part presentation; wrong part range for R&R.
    Machines Sensor drift; unstable warm-up; PLC timing issues.
    Materials Surface/finish effects ignored; non-representative parts chosen.
    Manpower Operator technique varies; no training or certification.
    Environment Temperature/humidity swings; ESD; vibration; contamination.


    Diagnostics & quick checks

    • Verify calibration status and confirm MSA plan exists for CTQs.
    • For variables: 10 parts × 3 operators × 2–3 repeats under production conditions.
    • For attributes: ≥50 randomized trials with repeats.
    • Use parts spanning process variation, not just nominal pieces.
    • Confirm adequate resolution — at least 1/10 of tolerance.
      Acceptance criteria
      Guidance from AIAG MSA (4th ed.) interprets %GRR < 10% as acceptable, 10–30% as application-dependent, and 30% as unacceptable. Number of distinct categories (ndc) should be ≥5, preferably ≥10. For attribute studies, aim
      for high repeatability and reproducibility, with overall % agreement and kappa in the ‘good’ range.


    30/60/90-day playbook
    0–30 days

    • List CTQs and link each to a gage/fixture.
    • Quarantine decisions relying on unqualified gages.
    • Run MSAs on top 3 CTQs.
    • Fix obvious issues — worn tips, poor fixtures, unstable environment.
      31–60 days
    • Train operators; create standard work for measurement methods.
    • Upgrade weak systems — better sensors, fixtures, environment controls.
    • Re-run MSAs to verify improvement.
    • Update control plans with qualified gages and SPC links.
      61–90 days
    • Institutionalize MSA re-studies after changes and at intervals.
    • Dashboard %GRR/ndc and attribute agreement by CTQ.
    • Feed MSA results into PFMEA detection rankings and LPA checks.
      Sustain & scale
      Sustain the gains by tying MSA to formal triggers: new gages, process changes, operator rotation, or SPC drift
      signals. Embed lessons into PFMEAs, control plans, and layered audits. When MSA is treated as a standing
      discipline, not a one-off event, data becomes trustworthy — and every decision downstream improves.

    • References & further reading
      [1] NIST/SEMATECH e-Handbook of Statistical Methods — Gauge R&R; Studies
      https://www.itl.nist.gov/div898/handbook/mpc/section4/mpc4.htm
      [2] ISO 22514-7:2021 — Capability of measurement processes
      https://www.iso.org/standard/80624.html
      [3] ASQ — Gage Repeatability and Reproducibility
      https://asq.org/quality-resources/gage-repeatability
      [4] SPC for Excel — Acceptance Criteria for MSA
      https://www.spcforexcel.com/knowledge/measurement-systems-analysis-gage-rr/acceptance-criteria-for-msa/

    📄 Download the full Deep Dive 01 PDF here