Welcome!

Government Cloud Authors: Elizabeth White, Liz McMillan, Gopala Krishna Behara, Raju Myadam, Kevin Jackson

Related Topics: @DevOpsSummit, Linux Containers, Containers Expo Blog

@DevOpsSummit: Blog Post

How to Stop Your Developers from Jumping Ship | @DevOpsSummit #Agile #DevOps

For developers working across a broken workflow, the frustration can be felt acutely

We are experiencing a ‘crisis of engagement,' according to a survey by management consultancy Gallup, with a staggering 87 percent of employees worldwide disengaged with their jobs. While there are many factors that can influence an employee's contentment - from support to compensation to job security - a common grievance in software development is disempowerment, where talented individuals are feeling hampered by their working environment.

Operating within a flawed system is infuriating in any profession. But for developers working across a broken workflow, the frustration can be felt more acutely. One of the major bones of contention is that developers find themselves working in a way that conflicts with their training and established best practices, one that is not conducive with Agile and DevOps working methodologies and prevents them from optimizing their tools-of-choice.

A main contributor to an unproductive working environment is a fragmented toolchain, which is a regular occurrence at organizations worldwide. Many of them are adopting a best-of-breed approach to tool procurement without a holistic investment strategy. Individual specialists, teams and disciplines are specifying a tool (or version thereof) as and when required - i.e., a project manager with CA Clarity PPM, a tester in HP ALM, a QA in HPE QC, a developer in JIRA and so on - with little consideration as to how it may impact the rest of the software lifecycle.

While it is logical to equip employees with the best tools for the job, this ad-hoc and unsynchronized approach is littered with time bombs that could explode at any time, negatively impacting a developer's satisfaction and, ultimately, an organization's software delivery capability. So what can be done to stop developers jumping ship to join Facebook, Netflix, et al?

Grounds for divorce
Any investment strategy must consider software lifecycle integration (SLI), otherwise an organization is quietly laying the foundation for an unproductive environment that will come back to haunt it. Tools in the software lifecycle are not designed to integrate with other third-party tools, meaning they do not naturally communicate with each other, and project-critical data typically becomes siloed within the tools.

The result is a workflow with little-to-no visibility, traceability or governance across the lifecycle, with stakeholders unable to share information with other stakeholders, greatly impacting collaboration. Such a dynamic can frustrate a developer in several ways, all of which are avoidable. Here are just a few of the issues caused by a disconnected toolchain:

  • Agile and DevOps blasphemy: Collaboration, interaction and information are the triumvirate of Agile and DevOps. The flow of artifacts between tools can be considered more important than the tool itself. As both methodologies are widely accepted as industry standards, it's understandable that most developers will be looking to apply the principles in their day-to-day work within their teams. But if their tool is isolated and they're unable to share or receive vital data at a click of button, they will become discontented as their training is wasted and skill set underutilized.
  • Developers just want to...develop: Developers are genuinely passionate about their work. They're problem solvers who love tackling issues with code, taking great pleasure in working with their colleagues to build and maintain projects, drive innovation, challenge the status quo and, significantly, create value for their company. But it's a tough job; intensive coding is an all-encompassing specialty that requires undivided attention. They don't - and shouldn't have to - spend valuable hours lost in boring non-value administrivia (status meetings, emails, data entry, spreadsheet merging, etc.). Such tasks can be minimalized, or even removed completely. If they're not, don't be surprised if your developers seek new pastures.
  • Friend or foe: If a tool isn't implemented correctly and working in unison with the lifecycle, it can quickly become a burden or, worse, a foe. Isolating a quality tool can quickly strip it of its capabilities and functionality, as developers spend more time fighting the tool than harnessing its powers. Benign issues become malignant, and great and engaged developers become frustrated and disinterested.
  • Rotten apple: As the adage goes, ‘one rotten apple spoils the whole bunch.' Team unity, especially in software development, is integral to Agile and DevOps principles. A strong software lifecycle needs all players in the game to be happy in their roles or balls will be dropped. If a developer is disenchanted by her role within the team because of inefficient processes and procedures, it's only a matter of time before her colleagues become demoralized too. Beware the domino effect.

Pain relief
Addressing a fragmented software lifecycle is the first step in creating harmony among your software development and delivery personnel. An integrated software value stream has many benefits, but at its heart, it's all about removing waste and getting rid of all the annoying work that plagues a developer's day.

What do we mean by ‘software value stream'? A value stream is a notion borrowed from Lean manufacturing; it's the sequence of activities an organization undertakes to deliver a customer request, focusing on both the production flow from raw material to end-product and the design flow from concept to realization.

Looking at software development from a value stream perspective puts the emphasis on creation of customer value, rather than simply looking at these activities as a process. It's all about the ‘Big Picture' - improving the whole process, not just the parts - to minimize waste and ensure customers get exactly what they asked for.

How does an integrated values stream boost employee engagement? The troubles experienced by a leading global bank during its four-year Agile and DevOps transformation illustrates the point perfectly. Given the nature of its business, it's critical that all information is consistent across all the bank's systems, yet it lacked an automated flow of information between tools.

This meant that developers were still rekeying from one tool to another, taking up to two hours a day in duplicate entry. Not only was this a huge waste of valuable labor hours that cost the business up to $10 million annually in lost value, developers often got so frustrated that they would leave. This case study is painfully familiar, and again, entirely avoidable.

A Developer's Paradise
An integrated software value stream yields many benefits for developers (and the rest of the software lifecycle), including:

  • Information flows automatically across teams without costly manual intervention and oversight, removing non-value work and bottlenecks.
  • Collaboration happens within the work, rather than in email or disconnected tools.
  • Reports and analytics emerge from a holistic view of all the artifacts.
  • Traceability and governance become a natural act rather than an expensive manual process.
  • Visibility into the value stream enables managers to understand project status and optimize processes, meaning developers have the freedom and insight to build extraordinary software.
  • A modular Agile toolchain means forward-thinking developers can plug in new tools (or new versions of their existing tools) to experiment with new ideas and innovations.

As a result, these organizations experience:

  • Faster time-to-market
  • Reduced development cost
  • Ability to add more features, more efficiently, while improving quality and reducing risk
  • Ability to leverage their software development capability to bring differentiating value to their businesses

And perhaps most importantly, it results in high developer engagement and job satisfaction; the fuel behind any software-driven organization that's looking to transform its business in a digital world.

More Stories By John Rauser

John Rauser is the IT Manager at Tasktop Technologies, a global enterprise software company. He also serves as VP Operations at the board of the Project Management Institute - Canadian West Coast Chapter, providing leadership and expertise on technology issues. He has a passion for discussing the business impacts of technology and analyzing strategies for managing IT.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@ThingsExpo Stories
DXWorldEXPO LLC, the producer of the world's most influential technology conferences and trade shows has announced the 22nd International CloudEXPO | DXWorldEXPO "Early Bird Registration" is now open. Register for Full Conference "Gold Pass" ▸ Here (Expo Hall ▸ Here)
As data explodes in quantity, importance and from new sources, the need for managing and protecting data residing across physical, virtual, and cloud environments grow with it. Managing data includes protecting it, indexing and classifying it for true, long-term management, compliance and E-Discovery. Commvault can ensure this with a single pane of glass solution – whether in a private cloud, a Service Provider delivered public cloud or a hybrid cloud environment – across the heterogeneous enter...
The Jevons Paradox suggests that when technological advances increase efficiency of a resource, it results in an overall increase in consumption. Writing on the increased use of coal as a result of technological improvements, 19th-century economist William Stanley Jevons found that these improvements led to the development of new ways to utilize coal. In his session at 19th Cloud Expo, Mark Thiele, Chief Strategy Officer for Apcera, compared the Jevons Paradox to modern-day enterprise IT, examin...
DXWorldEXPO LLC announced today that ICC-USA, a computer systems integrator and server manufacturing company focused on developing products and product appliances, will exhibit at the 22nd International CloudEXPO | DXWorldEXPO. DXWordEXPO New York 2018, colocated with CloudEXPO New York 2018 will be held November 11-13, 2018, in New York City. ICC is a computer systems integrator and server manufacturing company focused on developing products and product appliances to meet a wide range of ...
Headquartered in Plainsboro, NJ, Synametrics Technologies has provided IT professionals and computer systems developers since 1997. Based on the success of their initial product offerings (WinSQL and DeltaCopy), the company continues to create and hone innovative products that help its customers get more from their computer applications, databases and infrastructure. To date, over one million users around the world have chosen Synametrics solutions to help power their accelerated business or per...
Bill Schmarzo, author of "Big Data: Understanding How Data Powers Big Business" and "Big Data MBA: Driving Business Strategies with Data Science," is responsible for setting the strategy and defining the Big Data service offerings and capabilities for EMC Global Services Big Data Practice. As the CTO for the Big Data Practice, he is responsible for working with organizations to help them identify where and how to start their big data journeys. He's written several white papers, is an avid blogge...
We are seeing a major migration of enterprises applications to the cloud. As cloud and business use of real time applications accelerate, legacy networks are no longer able to architecturally support cloud adoption and deliver the performance and security required by highly distributed enterprises. These outdated solutions have become more costly and complicated to implement, install, manage, and maintain.SD-WAN offers unlimited capabilities for accessing the benefits of the cloud and Internet. ...
Charles Araujo is an industry analyst, internationally recognized authority on the Digital Enterprise and author of The Quantum Age of IT: Why Everything You Know About IT is About to Change. As Principal Analyst with Intellyx, he writes, speaks and advises organizations on how to navigate through this time of disruption. He is also the founder of The Institute for Digital Transformation and a sought after keynote speaker. He has been a regular contributor to both InformationWeek and CIO Insight...
Join IBM November 1 at 21st Cloud Expo at the Santa Clara Convention Center in Santa Clara, CA, and learn how IBM Watson can bring cognitive services and AI to intelligent, unmanned systems. Cognitive analysis impacts today’s systems with unparalleled ability that were previously available only to manned, back-end operations. Thanks to cloud processing, IBM Watson can bring cognitive services and AI to intelligent, unmanned systems. Imagine a robot vacuum that becomes your personal assistant tha...
"MobiDev is a software development company and we do complex, custom software development for everybody from entrepreneurs to large enterprises," explained Alan Winters, U.S. Head of Business Development at MobiDev, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
I think DevOps is now a rambunctious teenager - it's starting to get a mind of its own, wanting to get its own things but it still needs some adult supervision," explained Thomas Hooker, VP of marketing at CollabNet, in this SYS-CON.tv interview at DevOps Summit at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
Recently, WebRTC has a lot of eyes from market. The use cases of WebRTC are expanding - video chat, online education, online health care etc. Not only for human-to-human communication, but also IoT use cases such as machine to human use cases can be seen recently. One of the typical use-case is remote camera monitoring. With WebRTC, people can have interoperability and flexibility for deploying monitoring service. However, the benefit of WebRTC for IoT is not only its convenience and interopera...
Cloud-enabled transformation has evolved from cost saving measure to business innovation strategy -- one that combines the cloud with cognitive capabilities to drive market disruption. Learn how you can achieve the insight and agility you need to gain a competitive advantage. Industry-acclaimed CTO and cloud expert, Shankar Kalyana presents. Only the most exceptional IBMers are appointed with the rare distinction of IBM Fellow, the highest technical honor in the company. Shankar has also receive...
It is of utmost importance for the future success of WebRTC to ensure that interoperability is operational between web browsers and any WebRTC-compliant client. To be guaranteed as operational and effective, interoperability must be tested extensively by establishing WebRTC data and media connections between different web browsers running on different devices and operating systems. In his session at WebRTC Summit at @ThingsExpo, Dr. Alex Gouaillard, CEO and Founder of CoSMo Software, presented ...
WebRTC is great technology to build your own communication tools. It will be even more exciting experience it with advanced devices, such as a 360 Camera, 360 microphone, and a depth sensor camera. In his session at @ThingsExpo, Masashi Ganeko, a manager at INFOCOM Corporation, introduced two experimental projects from his team and what they learned from them. "Shotoku Tamago" uses the robot audition software HARK to track speakers in 360 video of a remote party. "Virtual Teleport" uses a multip...
Business professionals no longer wonder if they'll migrate to the cloud; it's now a matter of when. The cloud environment has proved to be a major force in transitioning to an agile business model that enables quick decisions and fast implementation that solidify customer relationships. And when the cloud is combined with the power of cognitive computing, it drives innovation and transformation that achieves astounding competitive advantage.
Data is the fuel that drives the machine learning algorithmic engines and ultimately provides the business value. In his session at Cloud Expo, Ed Featherston, a director and senior enterprise architect at Collaborative Consulting, discussed the key considerations around quality, volume, timeliness, and pedigree that must be dealt with in order to properly fuel that engine.
IoT is rapidly becoming mainstream as more and more investments are made into the platforms and technology. As this movement continues to expand and gain momentum it creates a massive wall of noise that can be difficult to sift through. Unfortunately, this inevitably makes IoT less approachable for people to get started with and can hamper efforts to integrate this key technology into your own portfolio. There are so many connected products already in place today with many hundreds more on the h...
When shopping for a new data processing platform for IoT solutions, many development teams want to be able to test-drive options before making a choice. Yet when evaluating an IoT solution, it’s simply not feasible to do so at scale with physical devices. Building a sensor simulator is the next best choice; however, generating a realistic simulation at very high TPS with ease of configurability is a formidable challenge. When dealing with multiple application or transport protocols, you would be...
Detecting internal user threats in the Big Data eco-system is challenging and cumbersome. Many organizations monitor internal usage of the Big Data eco-system using a set of alerts. This is not a scalable process given the increase in the number of alerts with the accelerating growth in data volume and user base. Organizations are increasingly leveraging machine learning to monitor only those data elements that are sensitive and critical, autonomously establish monitoring policies, and to detect...