The work balance chart is a critical component of a line balancing effort. It is both the graphical representation of the allocation of task time among operators, equipment, and transfers in a manufacturing or service process and a tool used to achieve an equal distribution.
Like other tools discussed in “The Third Degree,” a work balance chart may be referenced by other names in the myriad resources available. It is often called an operator balance chart, a valid moniker if only manual tasks are considered. It is also known as a Yamazumi Board. “Yamazumi” is Japanese for “stack up;” this term immediately makes sense when an example chart is seen, but requires an explanation to every non-Japanese speaker one encounters. Throughout the following presentation, “work balance chart,” or “WBC,” is used to refer to this tool and visual aid. This term is the most intuitive and characterizes the tool’s versatility in analyzing various forms of “work.”
0 Comments
A cause & effect diagram is best conceptualized as a specialized application and extension of an affinity diagram. Both types of diagram can be used for proactive (e.g. development planning) or reactive (e.g. problem-solving) purposes. Both use brainstorming techniques to collect information that is sorted into related groups. Where the two diverge is in the nature of relationships represented.
An affinity diagram may present several types of relationships among pieces of information collected. A cause & effect diagram, in contrast, is dedicated to a single relationship and its “direction,” namely, what is cause and what is effect. Most manufactured goods are produced and distributed to the marketplace where consumers are then sought. Services, in contrast, are not “produced” until there is a “consumer.” Simultaneous production and consumption is a hallmark of service; no inventory can be accumulated to compensate for fluctuating demand.
Instead, demand must be managed via predictable performance and efficiency. A service blueprint documents how a service is delivered, delineating customer actions and corresponding provider activity. Its pictorial format facilitates searches for improvements in current service delivery and identification of potential complementary offerings. A service blueprint can also be created proactively to optimize a delivery system before a service is made available to customers. To be effective in any pursuit, one must understand its objectives and influences. One influence, typically, has a greater impact on process performance than all others – the dominant characteristic of the process. The five main categories of process dominance are worker, setup, time, component, and information.
Processes require tools tailored to manage the dominant characteristic; this set of tools comprises a process control system. The levels of Operations Management at which the tools are employed, or the skills and responsibility for process performance reside, differ among the types of dominance. This installment of “The Third Degree” explores categories of process dominance, tools available to manage them, and examples of processes with each dominant characteristic. Responsibility for control of processes exhibiting each category of dominance will also be discussed in terms of the “Eight Analogical Levels of Operations Management.” Effective Operations Management requires multiple levels of analysis and monitoring. Each level is usually well-defined within an organization, though they may vary among organizations and industries. The size of an organization has a strong influence on the number of levels and the makeup and responsibilities of each.
In this installment of “The Third Degree,” one possible configuration of Operations Management levels is presented. To justify, or fully utilize, eight distinct levels of Operations Management, it is likely that an organization so configured is quite large. Therefore, the concepts presented should be applied to customize a configuration appropriate for a specific organization. Standards and guidelines published by industry groups or standards organizations typically undergo an extensive review process prior to acceptance. A number of drafts may be required to refine the content and format into a structure approved by a committee of decision-makers.
As one might expect, the draft review and approval process is not consistent for every publication. The number of drafts, time to review, and types of changes requested will vary. Though each review is intended to be rigorous, errors often remain in the approved publication. The content may also require interpretation to employ effectively. This is certainly true of the aligned AIAG/VDA FMEA Handbook. In this installment of the “FMEA” series, the Handbook’s errors and omissions, opacities and ambiguities will be discussed. Where possible, mistakes will be corrected, blanks filled in, and clarity provided in pursuit of greater utility of the Handbook for all FMEA practitioners. As mentioned in the introduction to the AIAG/VDA aligned standard (“Vol. V: Alignment”), the new FMEA Handbook, is a significant expansion of its predecessors. A substantial portion of this expansion is the introduction of a new FMEA type – the Supplemental FMEA for Monitoring and System Response (FMEA-MSR).
Modern vehicles contain a plethora of onboard diagnostic tools and driver aids. The FMEA-MSR is conducted to evaluate these tools for their ability to prevent or mitigate Effects of Failure during vehicle operation. Discussion of FMEA-MSR is devoid of comparisons to classical FMEA, as it has no correlate in that method. In this installment of the “FMEA” series, the new analysis will be presented in similar fashion to the previous aligned FMEA types. Understanding the aligned Design FMEA method is critical to successful implementation of FMEA-MSR; this presentation assumes the reader has attained sufficient competency in DFMEA. Even so, review of aligned DFMEA (Vol. VI) is highly recommended prior to pursuing FMEA-MSR. To conduct a Process FMEA according to AIAG/VDA alignment, the seven-step approach presented in Vol. VI (Aligned DFMEA) is used. The seven steps are repeated with a new focus of inquiry. Like the DFMEA, several system-, subsystem-, and component-level analyses may be required to fully understand a process.
Paralleling previous entries in the “FMEA” series, this installment presents the 7-step aligned approach applied to process analysis and the “Standard PFMEA Form Sheet.” Review of classical FMEA and aligned DFMEA is recommended prior to pursuing aligned PFMEA; familiarity with the seven steps, terminology used, and documentation formats will make aligned PFMEA more comprehensible. To differentiate it from “classical” FMEA, the result of the collaboration between AIAG (Automotive Industry Action Group) and VDA (Verband der Automobilindustrie) is called the “aligned” Failure Modes and Effects Analysis process. Using a seven-step approach, the aligned analysis incorporates significant work content that has typically been left on the periphery of FMEA training, though it is essential to effective analysis.
In this installment of the “FMEA” series, development of a Design FMEA is presented following the seven-step aligned process. Use of an aligned documentation format, the “Standard DFMEA Form Sheet,” is also demonstrated. In similar fashion to the classical DFMEA presentation of Vol. III, the content of each column of the form will be discussed in succession. Review of classical FMEA is recommended prior to attempting the aligned process to ensure a baseline understanding of FMEA terminology. Also, comparisons made between classical and aligned approaches will be more meaningful and, therefore, more helpful. Preparations for Process Failure Modes and Effects Analysis (Process FMEA) (see Vol. II) occur, in large part, while the Design FMEA undergoes revision to develop and assign Recommended Actions. An earlier start, while ostensibly desirable, may result in duplicated effort. As a design evolves, the processes required to support it also evolve; allowing a design to reach a sufficient level of maturity to minimize process redesign is an efficient approach to FMEA.
In this installment of the “FMEA” series, how to conduct a “classical” Process FMEA (PFMEA) is presented as a close parallel to that of DFMEA (Vol. III). Each is prepared as a standalone reference for those engaged in either activity, but reading both is recommended to maintain awareness of the interrelationship of analyses. In the context of Failure Modes and Effects Analysis (FMEA), “classical” refers to the techniques and formats that have been in use for many years, such as those presented in AIAG’s “FMEA Handbook” and other sources. Numerous variations of the document format are available for use. In this discussion, a recommended format is presented; one that facilitates a thorough, organized analysis.
Preparations for FMEA, discussed in Vol. II, are agnostic to the methodology and document format chosen; the inputs cited are applicable to any available. In this installment of the “FMEA” series, how to conduct a “classical” Design FMEA (DFMEA) is presented by explaining each column of the recommended form. Populating the form columns in the proper sequence is only an approximation of analysis, but it is a very useful one for gaining experience with the methodology. Failure Modes and Effects Analysis (FMEA) is most commonly used in product design and manufacturing contexts. However, it can also be helpful in other applications, such as administrative functions and service delivery. Each application context may require refinement of definitions and rating scales to provide maximum clarity, but the fundamentals remain the same.
Several standards have been published defining the structure and content of Failure Modes and Effects Analyses (FMEAs). Within these standards, there are often alternate formats presented for portions of the FMEA form; these may also change with subsequent revisions of each standard. Add to this variety the diversity of industry and customer-specific requirements. Those unbeholden to an industry-specific standard are free to adapt features of several to create a unique form for their own purposes. The freedom to customize results in a virtually limitless number of potential variants. Few potential FMEA variants are likely to have broad appeal, even among those unrestricted by customer requirements. This series aims to highlight the most practical formats available, encouraging a level of consistency among practitioners that maintains Failure Modes and Effects Analysis as a portable skill. Total conformity is not the goal; presenting perceived best practices is. Choosing effective strategies for waging war against error in manufacturing and service operations requires an understanding of “the enemy.” The types of error to be combatted, the sources of these errors, and the amount of error that will be tolerated are important components of a functional definition (see Vol. I for an introduction).
The traditional view is that the amount of error to be accepted is defined by the specification limits of each characteristic of interest. Exceeding the specified tolerance of any characteristic immediately transforms the process output from “good” to “bad.” This is a very restrictive and misleading point of view. Much greater insight is provided regarding product performance and customer satisfaction by loss functions. Myriad tools have been developed to aid collaboration of team members that are geographically separated. Temporally separated teams receive much less attention, despite this type of collaboration being paramount for success in many operations.
To achieve performance continuity in multi-shift operations, an effective pass-down process is required. Software is available to facilitate pass-down, but is not required for an effective process. The lowest-tech tools are often the best choices. A structured approach is the key to success – one that encourages participation, organization, and consistent execution. There is some disagreement among quality professionals whether or not precontrol is a form of statistical process control (SPC). Like many tools prescribed by the Shainin System, precontrol’s statistical sophistication is disguised by its simplicity. The attitude of many seems to be that if it isn’t difficult or complex, it must not be rigorous.
Despite its simplicity, precontrol provides an effective means of process monitoring with several advantages (compared to control charting), including:
Lesser known than Six Sigma, but no less valuable, the Shainin System is a structured program for problem solving, variation reduction, and quality improvement. While there are similarities between these two systems, some key characteristics lie in stark contrast.
This installment of “The War on Error” introduces the Shainin System, providing background information and a description of its structure. Some common problem-solving tools will also be described. Finally, a discussion of the relationship between the Shainin System and Six Sigma will be presented, allowing readers to evaluate the potential for implementation of each in their organizations. Despite the ubiquity of corporate Six Sigma programs and the intensity of their promotion, it is not uncommon for graduates to enter industry with little exposure and less understanding of their administration or purpose. Universities that offer Six Sigma instruction often do so as a separate certificate, unintegrated with any degree program. Students are often unaware of the availability or the value of such a certificate.
Upon entering industry, the tutelage of an invested and effective mentor is far from guaranteed. This can curtail entry-level employees’ ability to contribute to company objectives, or even to understand the conversations taking place around them. Without a structured introduction, these employees may struggle to succeed in their new workplace, while responsibility for failure is misplaced. This installment of “The War on Error” aims to provide an introduction sufficient to facilitate entry into a Six Sigma environment. May it also serve as a refresher for those seeking reentry after a career change or hiatus. While Vol. IV focused on variable gauge performance, this installment of “The War on Error” presents the study of attribute gauges. Requiring the judgment of human appraisers adds a layer of nuance to attribute assessment. Although we refer to attribute gauges, assessment may be made exclusively by the human senses. Thus, analysis of attribute gauges may be less intuitive or straightforward than that of their variable counterparts.
Conducting attribute gauge studies is similar to variable gauge R&R studies. The key difference is in data collection – rather than a continuum of numeric values, attributes are evaluated with respect to a small number of discrete categories. Categorization can be as simple as pass/fail; it may also involve grading a feature relative to a “stepped” scale. The scale could contain several gradations of color, transparency, or other visual characteristic. It could also be graded according to subjective assessments of fit or other performance characteristic. While you may have been hoping for rest and relaxation, the title actually refers to Gauge R&R – repeatability and reproducibility. Gauge R&R, or GRR, comprises a substantial share of the effort required by measurement system analysis. Preparation and execution of a GRR study can be resource-intensive; taking shortcuts, however, is ill-advised. The costs of accepting an unreliable measurement system are long-term and far in excess of the short-term inconvenience caused by a properly-conducted analysis.
The focus here is the evaluation of variable gauges. Prerequisites of a successful GRR study will be described and methodological alternatives will be defined. Finally, interpretation of results and acceptance criteria will be discussed. There is a “universal sequence for quality improvement,” according to the illustrious Joseph M. Juran, that defines the actions to be taken by any team to effect change. This includes teams pursuing error- and defect-reduction initiatives, variation reduction, or quality improvement by any other description.
Two of the seven steps of the universal sequence are “journeys” that the team must take to complete its problem-solving mission. The “diagnostic journey” and the “remedial journey” comprise the core of the problem-solving process and, thus, warrant particular attention. Of the “eight wastes of lean,” the impacts of defects may be the easiest to understand. Most find the need to rework or replace a defective part or repeat a faulty service, and the subsequent costs, to be intuitive. The consequences of excess inventory, motion, or transportation, however, may require a deeper understanding of operations management to fully appreciate.
Conceptually, poka yoke (poh-kah yoh-keh) is one of the simplest lean tools; at least it was at its inception. Over time, use of the term has morphed and expanded, increasing misuse and confusion. The desire to appear enlightened and lean has led many to misappropriate the term, applying it to any mechanism used, or attempt made, to reduce defects. Poka yoke is often conflated with other process control mechanisms, including engineering controls and management controls. To effectively reduce the occurrence of errors and resultant defects, it is imperative that process managers differentiate between poka yoke devices, engineering controls, and management controls. Understanding the capabilities and limitations of each allows appropriate actions to be taken to optimize the performance of any process. Every organization wants error to be kept at a minimum. The dedication to fulfilling this desire, however, often varies according to the severity of consequences that are likely to result. Manufacturers miss delivery dates or ship faulty product; service providers fail to satisfy customers or damage their property; militaries lose battles or cause civilian casualties; all increase the cost of operations.
You probably have some sensitivity to the effects errors have on your organization and its partners. This series explores strategies, tools, and related concepts to help you effectively combat error and its effects. This is your induction; welcome to The War on Error. Uses of augmented reality (AR) in various industries has been described in previous installments of “Augmented Reality” (Part 1, Part 2). In this installment, we will explore AR applications aimed at improving customer experiences in service operations. Whether creating new service options or improving delivery of existing services, AR has the potential to transform our interactions with service providers.
Front-office operations are mostly transparent due to customer participation. Customer presence is a key characteristic that differentiates services from the production of goods. Thus, technologies employed in service industries are often highly visible. This can be a blessing or a curse. Some of the augmented reality (AR) applications most likely to attract popular attention were presented in “Part 1: An Introduction to the Technology.” When employed by manufacturing companies, AR is less likely to be experienced directly by the masses, but may have a greater impact on their lives. There may be a shift, however, as AR applications pervade product development and end-user activities.
In this installment, we look at AR applications in manufacturing industries that improve operations, including product development, quality control, and maintenance. Some are involved directly in the transformation of materials to end products, while others fill supporting roles. The potential impact on customer satisfaction that AR use provides will also be explored. When we see or hear a reference to advanced technologies, many of us think of modern machinery used to perform physical processes, often without human intervention. CNC machining centers, robotic work cells, automated logistics systems, drones, and autonomous vehicles often eclipse other technologies in our visions. Digital tools are often overlooked simply because many of us find it difficult to visualize their use in the physical environments we regularly inhabit.
There is an entire class of digital tools that is rising in prominence, yet currently receives little attention in mainstream discourse: augmented reality (AR). There are valid applications of AR in varied industries. Increased awareness and understanding of these applications and the potential they possess for improving safety, quality, and productivity will help organizations identify opportunities to take the next step in digital transformation, building on predecessor technologies such as digital twins and virtual reality. |
AuthorIf you'd like to contribute to this blog, please email jay@jaywink.com with your suggestions. Archives
November 2023
Categories
All
![]() © JayWink Solutions, LLC
|