Discussions with clients and account teams often trigger thoughts. I’m trying to share those with you, looking for your feedback and comments. Here is another.
Product Lifecycle Management is slowly but surely gaining traction in discrete manufacturing, and is becoming the backbone of the product information in the company. As such it complements the ERP environment that contains the order information. PLM is supposed to track the product from initial conception to end of life, ERP the order from receipt to delivery 9and payment). The fact the two concepts have evolved separately does not make the interface any easier, but that’s not the topic I want to discuss today.
PLM is being implemented in many companies today, although in most situations it has more of a PDM (Product Data Management) flavour.
As companies increase the outsourcing of manufacturing, soon to be followed by the outsourcing of key design tasks, looking at expanding the reach of PLM beyond the boundaries of the company makes sense. But a number of issues need to be addressed:
Ownership of data, IP protection and the presence of potential competitors in the ecosystem require a segmentation of the data to ensure the confidentiality of the information entered by each of the participants
Geographical location of the players may result in having to prove the compliance with specific regulations such as ITAR.
Network latency influences information update, which may result in a system that is not easily accepted by some of the users.
So, to make a PLM environment truly usable across an ecosystem, a couple elements need to be taken into account. Data needs to be segregated ensuring each of the participants only has access to information he/she truly requires and is allowed to access. Data items may have to be duplicated to ensure appropriate latency, but in that case consistency is critical.
There are basically three approaches to achieve this, and I happen to know companies having taken all three, so there does not seem to be a right or wrong answer.
Build a large, integrated PLM environment that is accessible both from within the enterprise and by partners. Through access rights, each of the users has access to the information he/she is allowed to access. From a data management/data integrity and compliance perspective, this is the easiest approach, however, network latency might be an issue for partners located far away.
Build and deploy small PLM appliances that are located at each of the sites and synchronize the relevant data between the systems addresses the bandwidth issue, but on the other hand, may result in partners working with outdated information on some occasions. Here the integration of that PLM appliance with the companies PD&E environment becomes critical. Although there is a master/slave relationship between the main PLM environment and the appliances, the compliance issue needs to be looked at very carefully as some information is not allowed to be located in specific geographies.
Using a cloud based approach where the PLM information is updated from the PLM master and allowing, with appropriate security levels, the partners to address that environment could be a reasonable compromise. The network latency issue may be present, depending on how the cloud is structured.
The main issue is, however, not the data aspect, but the business processes and how those are executed across the ecosystem. But that will be covered in a next post.