As an emerging tech vendor, you know that it can be a struggle to use content marketing to create brand awareness. Researching and creating a compelling piece that will get the attention of your target audience takes a lot of research and time. Fortunately, there is an easy way to satisfy the need for content your audience craves: third-party content.

Third-party content is a form of content creation that exists in working with a credible, external source that will prepare a piece of thought-leadership for you. Let’s have a look at the advantages of third-party content and how to best use it.

Build Trust with Third-Party Content

Working with a well-known, established partner will not only give you ready-made content, but will also help build trust in your brand’s expertise, expand your reach, and raise awareness in a noisy market.

Did you know that on average, tech buyers download six pieces of content throughout the tech purchase process? With an increasing number of content and channels to choose from, they naturally prefer content from a trusted, well-known source. In fact, as many as 95% of tech buyers recommend vendors add more insight from industry thought leaders and analysts to improve their content.

Working with an established source like an analyst can create a piece of thought-leadership that elevates your brand’s position in the market and builds a foundation of trust and credibility.

Prepare, Prepare, Prepare

Third-party partners will work with you to create content that is closely aligned with your marketing objectives. To get the best out of your new content piece, you must do your homework before briefing your partner.

  • Which topic and keywords resonate most with your audience?
  • Which style of research will best serve you?
  • What’s the most compelling visualization that should accompany the piece of content?

Remember that you should keep in touch with your third-party representative during the creation of the piece. If your external partner has dedicated Success Managers, they will ensure you will have guidance from preparation to implementation.

Share

You might be tempted to take your new piece of content and publish it on all your marketing channels. This could lead to a spike in interest for your company. But it could also be the wrong channel for the message you are sending or even the wrong time (holiday season, anyone?).

We recommend you keep these tips in mind when marketing your third-party piece:

Strategy: don’t use your content at the same time on every channel. You could start by publishing it on your website and social media, use it in a blog post in the following months, then refer to it in a digital discussion on LinkedIn. The possibilities are endless. Our new eBook includes a handy checklist for your next marketing campaign.

Be Insightful

When you share your third-party content, you don’t want to simply post the content and add a generic message like “Good read!”.

Instead, you want to personalize your message to your audience and tell them why this is compelling information for them, and what the net gains from reading are. We are all busy and suffering from content overload and adding this bit of information makes you stand out from the crowd.

Avoid Marketing Pitfalls

With a little bit of savvy marketing, a great piece of third-party content will create awareness and credibility. But failing to establish some details before the launch of your campaign will quickly lead to a loss of ROI.

Make sure you

  • Are clear about your KPIs. What does success look like for this campaign? And for your company?
  • Be clear about your digital strategy and develop a concise marketing plan.
  • Track your success and refine constantly.

If you find these insights helpful and are looking for a partner, we might have the solution. IDC’s Thought Leadership Analyst Brief provides leading-edge third-party quality content that elevates your brand image and associates your company with emerging technology trends, driving your global media coverage and market awareness, and re-defining your client dialogue.

This year, the UNHCR announced that the number of people forcibly displaced exceeded 100 million globally for the first time on record. This means 1 in every 78 people has been forced to flee their homes. The number of people displaced, internally or externally, from weather-related disasters, famine and unemployment is also rising. The Institute for Economics and Peace (IEP) estimates that there could be 1.2 billion climate refugees by 2050.

A proactive approach is needed to help manage this perennial challenge and to improve the lives of refugees. Or, in the words of the Mayor of Warsaw, Poland, “it is time we phased out improvisation and instead created a strategy for coping and appropriate systems for helping refugees.”

To create more effective and empathetic systems for refugees, governments and their partners should consider taking a life-event approach. The life-event approach to digital service delivery includes bringing a range of services together to coincide with a particular event, such as a birth, marriage, and death. Rather than people having to reach out to a variety of agencies to access different services, governments can integrate information and resources to provide these services in a more seamless way.

More mature approaches also include service coordination and data exchange across agencies as well as with private sector and civil society organisations.

Some governments, including the US Federal Government, have started to take a life event approach to disaster response, from floods to wildfires. These sudden and unplanned life events require citizens and residents to access services from several agencies and require a more empathetic and personalised approach.

The same can be applied to the arrival of a refugee. When refugees arrive in a new country, they generally need to register for residency and digital identity, whether through the government or an NGO, and need immediate access to food, cash programs, and shelter. These short-term needs then morph into longer terms needs, including accessing healthcare services, education, and the labour market.

Taking a life event approach and bundling these services together can help to make often bureaucratic systems and services easier to navigate. This is even more important for vulnerable populations in a time of need and uncertainty. It can also improve efficiency, save time, and reduce costs for government agencies.

For example, in New Zealand, the government created SmartStart, a cross-agency online service to help parents navigate government services around the birth of a child. In its first year, the service resulted in 6,000 fewer visits to the Ministry of Social Development and has been well received by parents, midwives, and NGOs.

However, this should not be seen as a replacement for face-face support when needed. Some groups of refugees, such as unaccompanied minors or people with cognitive or physical disabilities, may require additional specialised and hands-on support from governments and NGOs

Source: IDC 2022

The Government of Portugal is leading the way in providing joined-up services for refugees. The Borders and Immigration Service has set up the Temporary Protection Regime for Ukrainian refugees. When Ukrainian refugees register through the online portal or in person, they receive identity numbers from key agencies including a tax identification number, social security identification number, and a national health service user number so that they can access key services.

Providing integrated services organised around planned or sudden lifetime events comes with its own challenges. It requires governments to set up the right governance mechanisms to collaborate across departments, share data in a secure and trusted way, and share budgets to enable integrated service delivery.

This is no small feat and requires a strategic approach and strong leadership, but there are examples of governments successfully overcoming these barriers.
National and local governments and international institutions looking to improve the lives and livelihoods of refugees should embrace this approach.

Leveraging technology for humanitarian purposes — HumTech — will require the right political motivation and leadership to turn refugee management from an ad hoc response to a crisis to a more systematic response to a perennial challenge.

For more information about technology and it’s impact on refugees:

To explore more of our coverage, please visit our Government Insights page.

Get in touch:

To learn more about Blockchain, Cryptocurrencies, NFTs and Web 3, how they are intertwined, and strategies to utilize these technologies for opportunities, read IDC’s new eBook, Blockchain, Crypto, NFTs, and Web3.

Web3 is no longer a hypothetical evolution of the current Web2.0, it is already taking shape and becoming reality as you read this. IDC defines Web3 as a collection of open technologies and protocols, including Blockchain (as Crypto and NFTs), that support the natively trusted use of decentralized data, knowledge, and value. In other words: Web3 will be built on a foundation of Crypto and NFTs, to equitably exchange value—both as currency and as content—between the creators of that content, the platforms which will host that content, and the end consumers of that content.

But what do we mean by “content”? Today, “content” is more than just someone’s cat pictures and cooking videos. In IDC’s definition, content is any and all data being shared from one entity to another: this can include cat pictures (which as we all know is the bedrock of social media), third party apps (such as games, i.e., what the original FarmVille had been), videos, music, stories, update posts, and similar. But it is also much, much more: data which is captured by content platforms every time a user watches a video, or “likes” a post, or spends a few seconds longer lingering on particular images, can be classified as content, as that user’s input, however minute, directly impacts the algorithms which determine what content will be shown more (or less) to other users with similar profiles.

All of this data has value from a business intelligence perspective, and therefore also monetary value, as this data, in Web3, will be more easily captured, stored, sorted, and repackaged as a marketable product, by utilizing NFTs (Non-Fungible Token). A single user’s entire Web history can be captured and stored, securely, on an NFT—a sort of long-term file of every uploaded picture, video, message, comment, click, like, view, and more—while on various pages and content platforms on the Web. In the ideal, utopian vision of Web3, that individual user would have total command and control over that data, stored in their NFT.

What is revolutionary is that through the use of NFTs, all this data and content can now be accurately and equitably tracked. The user actually creating content—whether it is a single person “liking” an image or a third-party app developer creating a game—can track and own their content, on an NFT. Now, if that content can be tracked, it means it can be purposefully exchanged for value. That value can range from a monetary renumeration (i.e., pay $4.99 a month to play the game, or pay $0.0001 for every “click” on social media platform) or access to benefits or perks (i.e., receive permission to access some of the user data from users playing your game, or gain access to VIP social media groups if you spend enough time engaging with posts, etc.). The critical factor is that by tracking all this in a secure, immutable, traceable, and unique record—an NFT—businesses can now offer to exchange value for that content and data.

THE DATA MARKETPLACE

Now that a business, such as a social media/content platform, can offer users monetary compensation (or special rights or access) in exchange for the right to use some, or all, of their content and data, it becomes much more appealing for users to willingly share their content and data. With an increase in user data, willingly shared by users, social media/content platforms can now repackage that data, and sell it on data marketplaces.

For example: A large fashion designer (i.e., LVMH) is planning their Spring catalogue for next year. They would like to know what are the hot trends that their target demographic are talking about or looking at this summer. So they go to a data marketplace, and place an open offer to buy a data set outlining what images or videos their target demographic have viewed. Let us say: people between the age of 22-30, living in major metropolitan areas (10+ million urban population), in the countries in which this fashion designer has stores, who viewed videos or pictures of clothing, during the summer months. The data, being packaged and sold (with permission of the users creating that data) in the data marketplace by either a social media platform (i.e., Facebook/Meta), a video content platform (i.e., TikTok or YouTube), or picture content platform (i.e., Instagram), can be priced fairly (with bid / ask pricing) and purchased by the fashion designer. Now, the data may bear out that videos and images featuring purple polka-dots get the most likes and views, and is trending higher and higher through the summer—leading them design a new product line for their spring catalogue with the soon to be hottest fashion of purple polka-dots.

CHANGING THE WAY BUSINESSES DO BUSINESS

In the above scenario, every participant is equitably compensated for their contribution, however small it may be (i.e., a “click” or “like” on an image), through the entire value creation process. The user has retained the right to sell their data/content, the platform company is being compensated for providing the content hosting platform and gathering and packaging the data, and the buyer of the data set (here, the fashion designer) is gaining value from having the most accurate and insightful data available to plan their business strategy. The foundational technology that allows for this accurate, secure, immutable, traceable, and unique record—and the ability to assign equitable value and compensation—is the NFT, and the underlying technology of the blockchain. Additionally, the rise in use of NFTs will coincide with the rise in use of cryptocurrencies. Forgive the pun, but NFTs are essentially the “other side of the coin” of cryptocurrencies. Cryptocurrencies permit the decentralized and secure exchange of digital value, whereas NFTs permit the decentralized and secure exchange of digital content.

Expect to see a rapid rise in the use of NFTs (beyond the hype of NFT art, which is an interesting proof of concept use case for NFTs), as businesses realize that any digital information—from cat pictures to music to pharmaceutical drug formulations—can be securely recorded and exchanged on an NFT. As NFTs rise, thus will cryptocurrencies, as the logical (and easier to use) exchange of digital value for digital content. To learn more about Blockchain, Cryptocurrencies, NFTs and Web 3, how they are intertwined, and strategies to utilize these technologies for opportunities, read IDC’s new eBook, Blockchain, Crypto, NFTs, and Web3. Click the button below to download the free eBook.

Let me present a modern problem: The bank sends me a new debit card, after the previous number had been exposed in a data breach. It’s great that they are protecting me and my money. It’s not so great that I now must update all of my subscriptions with my new card number. And I have a lot of subscriptions, especially for media. Prime. Fubo. Netflix. HBO Max. Showtime. Crunchy Roll. Peacock. And on and on and on.

So during a recent July evening, we sit down to enjoy a true crime documentary about a murder in New England (my favorite). It’s on one of my streaming providers, sold via a bundle with other services. I update the billing info on the parent service, go back to documentary and…no deal. I get a message to update my billing with the parent company of the bundle, which I had done. No problem, try again. The parent company showed I had indeed updated my billing info. Back to the bundled service for the murder show…and., once again, nothing. After about five tries, a light bulb went off.

These services were still using a high latency — probably batch — process to connect their systems and distribute data for billing and subscriptions. That means they were not syncing in real time. If real time is like sending a text message, their process was like sending a letter. 

Now as far as stakes go, this one was relatively low. The streaming service missed out on ad revenue, while annoying me. But what if the stakes were higher? What if I was hoping to catch a live, once-in-a-lifetime event? What if this was a service that kept my lights on? Or helped me get somewhere after my car broke down? What if I real-time was imperative to my customer satisfaction?

And now that I am saying this out loud (or typing it), I wonder if any customer interaction can truly be considered low stakes.  Losing out on ad revenue is not nothing. When it comes to media, I subscribe to way too many streaming services as it is. Did this negative, high latency interaction make me miss my old cable company? And in light of inflation concerns and tightening my household budget, a bad customer experience might have me thinking, “I can always find a true crime documentary about a New England murder on another streaming service. They are prolific.

Today, customers expect a real time experience. Not to sound like Veruca Salt, but we know what we want and want it now. High latency isn’t just a week or a day; it can also be minutes. Delays in data can put the business at risk.  

Business happens in real time. People buy things in real time, banking happens in real time, goods are shipped in real time, and bad players try to access your data in real time.

Enterprise intelligence means making decisions based on data, so don’t we want to use the freshest data available? Companies with higher levels of enterprise intelligence report greater revenue growth and customer acquisition. In fact, the top three goals realized by the intelligent enterprise are revenue growth, increase in innovation, and improvement in customer experience.

Considering these uncertain economic times, IDC recommends that companies invest in digital-first technologies to accelerate transformation and meet the headwinds of inflation. That includes investment in services that augment agility and enhance customer engagement. Real-time experiences can do that. In a recent survey, IT leaders told IDC that for 2022, investing in technology to achieve real time decision making is a top priority.

Change data capture (CDC) is a great example of implementing dynamic data movement for real-time results. CDC is very popular within the enterprise for populating their streaming data pipelines, and there are new offerings in this space overt the past 12 months to accommodate a growing list of data sources and sinks.

CDC is just one entry point into real time streaming data. And not every use case requires zero-latency solutions. But we all know that improving the customer experience (CX) directly impacts the organization’s bottom line. With phones in our hands and digital assistants in our kitchens, we want relevant experiences that satisfy our needs with the least effort.

Streaming data use cases exist across all industries, from manufacturing to financial, from retail to healthcare. Our research shows that strategic use of streaming data across departments aligns with higher levels of digital maturity, i.e. the more mature organizations are implementing a streaming strategy across departments and throughout the enterprise, while the more immature organizations are still siloed in their use case implementations. When business runs in real-time, scheduled batch data is often out-of-date, and use cases we see around security/threat management, customer activity tracking, and real-time financial data plays that theory out.

I always say, “Let the use case lead.” It should direct how you think about real-time architecture, which ideally, is an expansion of your existing framework to avoid creating data silos. The top questions to ask are: “What are the business benefits we are hoping to achieve?” and “What insights do we need to achieve those goals?” But it’s also important to consider how a use case can drive insights at scale by asking: “Who needs those insights and where do they need them?” and “What other systems might we need to integrate with for context or to operationalize insights?” Asking these questions will help your company figure out the technology requirements, as well as where you need to be able to deploy those capabilities and what applications and systems you need to connect.

Amy Machado - Sr. Research Manager - IDC

Amy Machado is a Senior Research Manager for IDC's Content and Knowledge Discovery Strategies program. She examines the evolution of the content and unstructured data supply chain and its impact on mission-critical operations. Driven by demand for using data to make business decisions and building AI-ready data sets, modern technologies are creating new opportunities for automation and insights for what has traditionally been locked in undefined unstructured data. Ms. Machado's coverage includes enterprise content management, content sharing and collaboration, knowledge discovery, knowledge management, intelligent document processing, and customer communications management software. She uses primary research to size, forecast, and segment these application markets. Ms. Machado has over 20 years of analyst experience and has been at IDC since 2011. In addition to content and knowledge services, she has also focused on real-time streaming data and time series analytics, production printing, and industrial printing markets.

In this session of the Digital Leaders Community, we had experts from Nutanix and IDC Metri on hand to share their experience with the community.

The interactive session started as usual with a roundup of the reasons attendees from the community had joined. There was a variety of interests among the digital leaders from around Europe including one company that has made the decision to move to the cloud, some that have plans based on the maturity of different workloads, and others that wanted to hear about peers that have already moved to cloud.

One public organisation told how it had become completely cloud based. Did it save money? The attendee commented on the need to make sure the CFO and finance team understood that opex would increase. There is a need to make sure finance sees the savings as well. They talked about the datacentre no longer being needed, so that space can be used as offices. It was concluded that FinOps is tricky, but especially important as it is hard to forecast costs as more people move to cloud.

Nutanix has many customers moving to cloud, and sees how cloud costs have evolved for clients. The expert from Nutanix, Steen Dalgas, in particular has been helping companies figure out the TCO of cloud and other environments. Cloud is great for companies in innovation/growth mode — to rent capacity when you don’t know what parts will work. It is better to keep mature areas on-premises, as it is much more cost effective.

You can compare with on-premises with infrastructure as a service (IaaS). Platform as a service (PaaS) is quite different, and is hard to compare with on-premises. It is also extremely hard to repatriate from PaaS. The majority of cloud use now is IaaS, but PaaS will start to become more dominant.

A word of caution — the attendees mostly agreed that “By moving to the cloud, you don’t get rid of problems, you just move them somewhere else.

There followed a discussion on cloud-related skills. One attendee noted that they are outsourcing the skills they need to core suppliers. Some people feel more secure in the cloud than on-premises. However, cloud provides an opportunity for people to be lazy — sloppy code with data leaks was the example — that is a problem on-premises that in the cloud can become very expensive very quickly.

There is major demand for FinOps — managing the different environments, figuring which workloads to use in which environment. And this is a key skill along with the software skill of internal cloud governance; what users buy from the cloud on company credit cards can be significant if they don’t understand the issues in terms of variable costs and security.

The meeting discussed whether you could prohibit people from using the cloud, or do you educate people about costs and let them get on with their work? You need to monitor cloud bills very closely, to educate people, and you need people with knowledge of cloud biz models. There is always a need to ask many questions about the cloud invoice and figure out what’s happening internally.

One suggestion was that line of business people need to be aware that what they do costs money. Ideally, make sure costs come straight back to users. This can be done by tagging cloud resources with an app name, environment name, or resource tag so you know who is using what and how much it costs. One comment was that unfortunately sometimes even IT doesn’t know how the cloud business model works. ­­

 

Two years ago, one global bank said it would move out of datacentres. Now it says it will keep DCs in major global centres but make them a cloud-like experience. One attendee related how their company tried a shared private cloud, but consider it as on-premises. They can control a lot, but don’t need to pay for the whole thing when capacity is not used.

The IDC CIO Advisory team would like to thank everyone who came to the call for their input. It is always inspiring to hear from those making changes in their business and taking the tough calls with their management colleagues. We hope this session was valuable and provided many takeaways for you.

If you already receive invitations to our sessions, I hope to see you there. If you would like to join this community, please email Marc Dowd mdowd@idc.com.

 

IDC Digital Leadership Community: Topics 3Q22

Brainstorming Out of the Box

Thursday, August 25, 2022, 17:00 CET

Sometimes doing something unexpected can really pay off. This is true of being a digital leader, just as it is for disruptive innovation.

In this light-hearted but serious session we will look at the “hacks” and experiences that we collectively can share. We will be dealing with subjects in all areas from leadership “masterstrokes we have witnessed” to unusual policies that work:

In this session you can expect:

  • Amusing anecdotes with a deeper meaning
  • Car crash policies and how to avoid them
  • The heroes of digital leadership, and why we think they are great

Data and Analytics/Data Culture/Data Management Technology

Thursday, September 29, 2022, 17:00 CET

I for one thought it would never happen: IDC research shows that organisations are looking to move to being data-driven. I see movements among my clients to gather all the data in the organisation that is relevant and make it available.

Is the dream coming true? Are we moving to a world where data will at last come into its own and become the “new oil”?

In this session, we will discuss how to move towards a better use of data. Topics will include:

  • Developing trust in data — is that the key?
  • Governance of data
  • Managing the complexity of data initiative
  • The security and sovereignty issues around the use of data

We hope you will join us. Please feel free to suggest digital leaders who might benefit from these discussions.

https://www.idc.com/eu/digital-leadership-advisory

Marc Dowd - Principal, Client Advisory - Research and Consulting - IDC

Marc Dowd is the principal for IDC’s European client advisory practice. Dowd has over 25 years of experience working with the leaders of corporate IT across a wide range of industries. This includes 9 years as principal for EMEA advising CIOs of large international companies and government bodies for Forrester Research. Recently he has been focusing on Digital Transformation (DX) and the use of emerging technology such as AI, IoT and blockchain to develop new business models and business capabilities. His experience enables him to provide CIOs and strategic business planners within organisations who use technology, with market and customer insight, analysis, tactical advice, forecasting and technology trend intelligence to senior management teams at local, regional and worldwide levels.

Let’s start with a question.

When was the last time you picked up a phone to call a coworker?

Odds are, it was a while ago.  Instead, you reached out through Teams or Slack, email or a workflow of some kind.  You might have pushed a task to them through a service portal or used an analytics tool to give them data.  And you probably didn’t think anything of it.  Maybe you used Dall-E to create a custom meme then folded it into fluid document.

Each of these scenarios, and many more, are components of what IDC calls the “Intelligent Digital Workspace” (IDW).  Well, we actually call it an “intelligent digital workspace ecosystem”, which is a phrase intended to convey:

  • Intelligent – the environment uses AI/ML and scripted automation to adapt to the user’s requirements
  • Digital – the environment coordinates electronics and software both local to the user and remote to them to meet the user’s needs
  • Workspace – the environment is used to do work, as opposed to, say, playing games
  • Ecosystem – the environment is deliberately constructed of many components (data, electronics, software, and more) which work together in a coordinated fashion

If we ask people directly if they have deployed an “intelligent digital workspace ecosystem”, not many respond positively.  But, if instead we look at the functions of such an ecosystem, we see that over 70% of companies have enough functions to say they have already deployed it.

How did we get here?

So, how did we end up deploying an entirely new operational concept of the workspace without, well, noticing?  Did vendors trick us into it?  Was it the result of a Friday night that we just, collectively, don’t quite remember but photos ended up online?

Honestly, like any technical ecosystem, the IDW concept emerged in response to the collective efforts of tens of thousands of companies to provide employees relief from:

As we all struggled to adapt to these challenges, we also discovered that those inefficient processes and the technical challenges with the applications supporting those processes spurred further issues.  Teams became increasingly isolated within themselves.  IT support simply could not respond quickly enough and small problems lingered for hours or days.  Security systems designed for on-premise use, or to protect systems for occasional remote work, were untenably obscure and difficult to use over an extended period of time.

So, of course, we fixed it.

We deployed management tools which helped smooth out end user security (1st ranked IDW management tool) and operations.  These tools came from major vendors and small players, grew up in-house and evolved quickly to meet our expanded needs.

We deployed team collaboration and workflow tools to connect employees (1st and 2nd ranked IDW end-user argumentation). 

We deployed intelligent service management to provide support (2nd ranked workspace integration) with cross-group collaboration and the ability to report on activity across teams and the enterprise being the two most important attributes we designed for.

When it wasn’t enough, we empowered our teams to develop their own solutions and deployed those enterprise wide, accomplishing more for “agile transformation” in two years than we did in the previous two decades.

What’s Next?

Although we did well, solving those problems brought with it a host of new opportunities.  The results are often extremely complex, interweaving systems never really intended to work together.  Each enterprise has to manage the complexity on their own, and frankly they seem tired of it.  Now over 75% of organizations say that management and maintenance of this elaborate ecosystem is the responsibility of the vendors not the enterprise. 

This is, in part, because the world does not stand still.  The seemingly impossible challenges of allowing people to work in a dispersed, hybrid environment have given way in the hierarchy of problems to questions of geopolitical instability, to a collapsing return on energy investment for fossil fuels, and labor-shortages/job shuffling across a broad demographic range.  These pressures will only continue – with storms of disruption sweeping through the technology industry.  Simply put, enterprises will expect their vendors to step up and manage these ecosystems because the IT staff will have other things to do.

Meeting these “other things” will require the enterprise to clarify how it intends to use the IDW and for chosen platform vendors to respond, allowing employees for all enterprises to deliver faster, more consistently, and with greater productivity than traditionally organized and supported teams.

Now What?

So, that’s a lot.  We de facto created a new kind of digital workspace, founded in a wide range of technologies ranging from virtualization and endpoint management on one side to enterprise portals and content management on the other.  We cannot spend the time to manage the complexity of it anymore, so will turn to outside vendors and AI/ML to reduce operational load.  So, what do we do about it?

The question isn’t if you have an IDW; it’s whether you are actively coordinating it to deliver the right outcomes.

Advice for the enterprise: Enterprises need to step back and assess their existing environments.  Many will find that they have, de facto, deployed elements of an intelligent digital workspace as part of their ongoing efforts.  Those that have deployed the IDW de facto need to rapidly fill in any missing capabilities and work with their existing vendors to get out from under the operational burden the new environment creates.  Those that for whatever reason, lagged behind need to move quickly, coordinating with new or emerging vendors to deploy enough capabilities to support new ways of working.

Advice for the vendors: Vendors need to stop and assess their current offerings in terms of where they fit into the intelligent digital workspace ecosystem.  Coordination with partners, the ability to quickly solve emerging problems, and lowering operational burdens are likely to be the key messages customers need to hear over the next 12 to 18 months.  Many companies will also look for AI/ML enabled support and operations services, as well as the ability to deploy AI/ML models to support their employees.

The IDW is now a permanent feature of the enterprise.  It will, whether we think about it or not, continue to expand.  By being intentional and focused, enterprises and their vendor partners can derive the best value from this evolving ecosystem, shaping it to their needs.

Shannon Kalvar - Research Director - IDC

Shannon Kalvar is Research Director for Enterprise Systems Management, Enterprise Client Platforms, Observability, and AIOps at IDC. His current research addresses the evolution of digital work, AI in both an end-user and IT operational contexts, and the transformation of IT management in a rapidly changing world where humans must master the exponential increase in data and complexity while navigating tumultuous economic and geopolitical times. Shannon is a member of the Future of Work and AI councils, as well as a contributor to IDCs quantum computing research. He is based in Indianapolis, IN.

Many senior managers who have gone through a transformation from a traditional software development organization to an organization with self-organizing Agile teams, understand that this was a necessary exercise to ensure that value is delivered faster to their everchanging business environment. Where the survival of companies is often a direct result of delivering new user functionality faster than the competition, these managers are usually happy they went through this, but they also recognize that adopting an Agile way of working is not a silver bullet or a guarantee for success.

Although the concept of self-organizing teams sounds great, they still need to be managed in the context of the entire organization, its mission, vision and strategy.

Senior management needs to make decisions regarding budgets, investments, staffing, what to do, and what not to do. They need to decide on the strategic directions, the targets and goals, the operating model but also monitor progress and make corrections when necessary.

What we see in the market is that organizations struggle with the latter, and this is largely due to a lack of objective management information. Many Agile teams use progress metrics that make sense in their team, but which can’t be aggregated into useful management information. However, it’s important to understand the performance of the Agile teams, as they are providing the value to the business and its customers. We see that companies operate on different levels of maturity when it comes to the control they have over their application development teams. The question is: on which maturity level is your company?

A well-known model to look at maturity of processes is the Capability Maturity Model Integration (CMMi) model.  This model assesses the development processes on 5 different levels (see figure).

CMMi maturity levels
Source: https://en.wikipedia.org/wiki/Capability_Maturity_Model_Integration

This is a good principle and can be used to assess control over Agile development teams as well. The Agile Control Maturity Model (ACMM), based on the CMMi levels, allows senior management to understand the control level regarding their agile teams and the value they deliver. The following table shows the way IDC Metri looks at the different levels of Agile control maturity.

Low maturity levels (0, 1 and 2)

Most organizations currently are on level 0 or 1. They went through the transformation, and CIO’s and other senior management got advice from Agile coaches and other consultants on how they should manage the new organization where self-organizing teams have a lot of power to decide, and the business is very closely working with the teams to create value. However, on level 0 or 1, you have no idea how much value is created for the budget spent, and how this relates to the competition. Furthermore, predictability is low and it’s hard to manage the interdependencies between teams and to forecast when certain pieces of software will be ready. External teams are contracted based on time and materials (hourly rates) and not output-based (maximum value for money).

High maturity levels (3, 4 and 5)

As an organization gets to a higher maturity level, senior management gets more and more in control. Objective metrics are displayed in up-to-date dashboards, enabling them to make decisions based on facts. This brings along many benefits, like for instance:

  • The ability to demonstrate to stakeholders that you (the CIO and Senior Manager) is in control of the performance and value creation of the Agile teams.
  • Focus on actual value creation by improving productivity and delivery speed, while improving the quality of the product.
  • Objective measurements are used to understand which are the high performing teams and which are the low performing teams. Then the reasons this can be investigated and improvement actions can be carried out.
  • Software Cost Estimation accuracy improves enormously, as data from the  organization and teams can be used to calibrate the parametric models.
  • Leading indicators result in increased predictability of the teams, eliminating schedule or cost slippages.
  • Based on objective measurements and targeted removal of critical violations, the risk in the application portfolio reduces significantly while quality, maintainability and cost levels improve.
  • Output-based metrics can be used in contracting Agile teams, enabling organizations to select the team that provides the most value for money (productivity times average rate) instead of the cheapest one that may not be productive at all.

And there are many more benefits.

In the next figure, typical productivity improvements we observed are shown, where 0% is the market average for comparable teams/applications.

Being an expensive function, Agile development teams should be managed to ensure value is optimized for the investment. But, how do you measure the value? When IT leaders are struggling with this common challenge, IDC Metri has prepared this checklist to help you manage and measure your teams’ performance.

IDC’s Digital Leadership Community (DLC) July 2022 session was much more positive in outlook than I had expected. The participants shared their experiences and plans openly, and it was quite remarkable to be part of a discussion that really focused on “war stories” from the field and great suggestions for action plans.

Inflation is making the war for talent more difficult and is impacting salary expectations for new hires. One member spoke of their finance department telling them that when the recession comes, salaries will fall. But will that really happen? One CIO told us it has already happened in Israel, where a crash in the start-up market has eased the situation from just months ago when requests for salary increases were in the order of 30%–50%. This could be a sign that the overheating in the salary space will subside somewhat due to recessionary pressures.

Nevertheless it was acknowledged that IT skills are in high demand and will remain so. This is reinforced by IDC research that says the gap in available skills is still growing in all markets. The discrepancy in salary between new hires and existing employees is widening due to the need to pay new hires more due to inflation, which is a potential source of friction between employees. One company in a market with inflation running very high said it has resorted to salary reviews every six months to stem the tide and ease the suffering of its IT workforce. Proactively offering wage increases to retain IT talent was discussed.

There was a consensus that inflation is hitting opex contracts hardest, as might be expected, but there were also warnings from participants that negotiations across the board are becoming harder as suppliers brace for rising costs. Many supplier companies are trying to raise prices to adjust to higher costs (including energy costs). One CIO spoke of an SD-WAN contract negotiation where he saw clauses in the contract to ensure flexibility in prices between now and when the contract is actually signed. It was the first time he’d seen such a clause. The consensus was “be ready to walk away from key suppliers if they continue to hold a gun to your head.”

All participants agreed they are expecting more such difficult discussions with vendors until the end of the year. These discussions will have a big impact on IT budgeting in October. For some, who have had contracts in place for more than 10 years, vendors are now asking for additional clauses to provide flexibility for prices.

A positive note was raised by a CIO from Turkey, which is currently at 78% inflation. He said they are used to inflation. “It makes us flexible,” he said. They are focusing on fixed-fee agreements and projects to try to push inflationary risk on to suppliers. They are also migrating to cheaper or free products — they used MS SQL previously, for example, and are now moving to NoSQL and alternatives. He also said, “Big inflation rates make the business side dynamic.” His organisation is consolidating IT across the international group and is planning to have most IT handled from Turkey, as salaries are quite low there, and this reduces costs in opcos in other countries. Essentially, for that CIO, inflation is a challenge but also an opportunity.

One attendee — who had just finished budgeting — said their budget had been agreed with less discussion than ever because the finance department knows that inflation is raging. For now, the company can pass on prices to the market. As a distributor you can benefit from inflation. IT has even provided a special calculation tool to optimise pricing for the distribution business unit. It was noted by participants that price increases are even more drastic due to the dollar-euro exchange rate changes with the strengthening of the dollar.

Attendees also discussed how to decide which projects to postpone. A number of participants said they are looking for higher ROI than usual, in effect taking a harder look than usual, with IT deciding together with business which projects would pay back quickly. An attendee said his organisation is keeping IT “on prem” to delay the shift to opex. Others have moved to outsourcing to save money by avoiding upfront investment.

Will there be a big drop in IT budgets with the expected recession? No one on the call was expecting a big drop. The recommendation from attendees was to slow down strategic projects that have a longer payback period (three to five years) and most have already done that. I asked if anyone expects budgeting to become more dynamic/frequent. The answer from the group was, “No, not really.” One attendee did note that their cloud/FinOps team is looking to forecast cloud costs more frequently to avoid cost surprises and suggested looking for support from AWS and MS for that. Some suggested conducting this assessment on a monthly basis, and getting finance team support for the IT team to do that. The need to educate the executive board about the volatility of cloud costs was also noted.

Energy prices are a crucial factor for some participants. In certain industries the business is doing scenario-based planning based on different exchange rate outcomes and energy price thresholds.

The meeting ended with surprising positivity among the attendees. It was good to see that the Digital Leadership Community is pulling together to help each other in times of adversity. We look forward to you joining us in our next meeting. If you would like to access the DLC board for this quarter and add comments or read the IDC research, please do so here.

If you already receive invitations to our sessions, I hope to see you there. If you would like to join the community, please email Marc Dowd (mdowd@idc.com).

 

IDC Digital Leadership Community – Think Tank Topic August 2022

Brainstorming Out of the Box

Thursday, August 25, 2022, 17:00 CET

Sometimes doing something unexpected can really pay off. This is true of being a Digital Leader, just as it is for disruptive innovation.

In this light-hearted but also serious session we will look at the “hacks” and experiences that we can share. We will be dealing with subjects in all areas from leadership “masterstrokes we have witnessed” to unusual policies that work.

In the session you can expect:

  • Amusing anecdotes with a deeper meaning
  • Car crash policies and how to avoid them
  • The heroes of digital leadership, and why we think they are great

We hope you will join us. Please feel free to suggest Digital Leaders who might benefit from these discussions.

https://www.idc.com/eu/digital-leadership-advisory

News about data breaches – when enterprises ‘share’ their data without consent – make headlines and strike fear into executives’ hearts, send IT departments into overdrive and shake customer and employee confidence. However, little is said about when enterprises share data purposefully. There is a strong case to be made about how sharing data can make you and your ecosystem partners stronger, more resilient to volatility and uncertainty, and better able to serve your customers.

Did you know that 50% of US and Canadian of enterprises are already sourcing data from their partners (suppliers, distributors, retailers, etc.) and an additional 19% say that they are likely to begin doing so in the near future? (See material from IDC’s March 2022 Data Buyers Study) This means that it is highly likely that your competition is doing this. It is not just that so many others are doing this, it is that there can be a lot of benefit to sharing with trusted partners.

Shared data and insight is one of the foundations of successful industry ecosystems. IDC’s most recent global research finds that more than 1/3 of enterprises find that shared data positively impacts KPIs for innovation rates, business agility, and transforming their individual enterprise to more digital operation (IDC’s Future of Industry Ecosystem Survey, June 2022). This type of sharing is widespread and can take many forms. There are obvious opportunities for suppliers to share data with the companies that make finished goods, for example. When the finished goods company has accurate information about supplies and transit times for inputs to their products, they can foresee potential out of stock situations. This can allow them to qualify additional suppliers and/or inform their customers of potential bottlenecks. Sourcing and sharing data outside the individual enterprise helps provide insight that one entity cannot effectively collect and analyze on its own.

Of course, after you answer, ‘Why share data and insight?’ you have to figure out, ‘How?’ IDC has been looking at this through a series of research, starting with The Why and the How of Shared Data in Industry Ecosystems where we describe ways to create value from data and insights through fit-for-purpose solutions such as data marketplaces. Next, we followed that up with case studies of how groups have made this work in a presentation on Shared Industry Ecosystem Data: How Ecosystems Made It Work with Data Sharing and Marketplace Technology. In this research, you can learn the types of technologies that others have used to make data and insight sharing more available. These tools also assist with curation, governance and entitlement – which are key to any shared data situation.

Many enterprises still struggle with the ability to share data and insight across divisions and geographies. However, you don’t have to wait until you get all of your internal information sharing perfect before you look to how outside collaboration might provide advantages. Wins with external stakeholders could pave the way for more effective collaboration internally!

Data sharing has existed since before there were technologies in place. Vendors and customers have often had dialog where they shared information, whether these arrangements were formal or informal. Most often, solving the technical side of data sharing meets with less skepticism than the governance, change management and pure inertia issues. The gap has been less that potential ecosystem participants are unable to share and more that they have been unwilling for one reason or another. IDC receives a lot of inquiries about topics such as:

  • What if there’s no industry ecosystem for me?
  • How do I know what to share?
  • How am I going to share this?
  • Do I have to share everything?
  • How do I get others to share with me?
  • Should I charge for access to my data?
  • Can we make money doing this?

Your enterprise is not alone in wanting to learn more but being cautious about the potential exposure of shared data and insights. And, even though a few ecosystems have been able to create an outside revenue stream for these outputs, it should not be a primary consideration when potential partners come together.

Reach out and let’s talk about how to make shared data and insight in industry ecosystems work for you! There a quite a few examples we can point to, across industries and across the globe. You can learn from other ecosystems and how they’ve used technology, process and teams to derive mutual benefit.

Next on our research agenda, we are crafting a TechBrief on Data Clean Rooms – a way to share data while preserving privacy and competitive information, while benefiting the two or more parties who are looking to tackle common challenges ranging from fighting fraud to improving supply chain issues to helping to gain more insight about customers and potential customers. This is a way to be building your own ecosystem and gaining valuable insight while keeping the legal and security teams satisfied with the level of care you put in place.

We will continue to write about the topic of shared data & insights within our Future of Industry Ecosystems practice as well as our Data as a Service practice.

Lynne Schneider - Research Director - IDC

Lynne Schneider is Research Director leading IDC's Data Collaboration & Monetization, and Location & Geospatial Intelligence market research and advisory practices. Ms. Schneider's core research coverage in DaaS includes data sourcing and delivery services from traditional and emerging data providers along with evolving data aggregation and dissemination platforms. The breadth of coverage includes services that enable an organization to externally monetize data generated as part of the organization's ongoing operations, value-added information derived from this data, and the marketplace for combining data with other solutions. This research analyzes the supply and demand side business and technology trends of this emerging category.

Organizations overcome challenges in quantifying Agile value by embracing a solution that assesses, benchmarks and course-corrects Agile development teams. How is this done? Through agile value management.

Agile development is challenging to manage and measure, especially when compared with traditional development models like Waterfall

Management is often uncertain when functionality will be delivered and at what cost, whether it delivers required quality, and with what inherent risks. Due to its inherent nature of using Story Points to assess progress, Agile value is challenging to quantify.

If something’s difficult to measure, it’s hard to evaluate, manage and ensure value delivery. Through an example, this blog sets out to show that it’s possible to counter these challenges and restore balance between the relationship of Agile teams and management.

Imagine this common scenario: A major client-facing application that digitally transforms the company is a year into development with delivery postponed every quarter. Sprints are chewing through the backlog, but it keeps re-filling with refactoring, bugs and newly discovered requirements. Turn-over on the team is creating challenges in productivity and quality, and onboarding new team members takes too long, which impacts overall productivity. 

Management keeps receiving progress planning updates that show a smooth, Agile development machine with sprints and stories paced to deliver, but then the budget keeps increasing and the delivery date shifts to the right. Management is struggling to reconcile the reports versus actual cost, productivity, quality, and deadline progress. They’ve replaced the project manager once, but are frustrated by conflicting messages and are distrustful of quality, timely progress. 

The development team is confused and team members are leaving the project (and the company) due to management pressure to correct a project, where the team feels it is moving through the Agile process effectively. They feel that “management” doesn’t understand how Agile works and its value proposition; coming from Waterfall project management, they believe that management simply doesn’t get Agile. 

The development team also feels hampered by staff turnover, resulting in lost productivity, institutional memory and decreased quality. If management got off their backs and let them work, they could deliver a successful product. 

Behind management’s interest in seeing the product done and delivered is that they see it as key to their digital transformation, and a competitive and defensive necessity versus their market peers. If Agile is so much better than Waterfall, why are costs, duration and quality spiraling out of original forecasts? At least with Waterfall you knew what the milestones were and could prioritize deliverables. With Agile, it’s all about backlog and Story Points. 

Management eventually lost faith in the development process to deliver the application in the competitive timeframe needed with a minimum viable product and a semblance of cost controls. They feel like they’ve lost grip over a project that has become out of hand with no end in sight. Both groups need a concrete measurable process that allows all parties to objectively and consistently assess the team performance, product health and quality. They need a dependable way to visualize progress (productivity, cost, delivery speed, quality, security, maintainability). 

The company previously subscribed to research from IDC and heard about the IDC Metri Agile Value Management (AVM) solution that assesses, benchmarks and course-corrects Agile development teams. AVM is a systematic, quantitative, comprehensive, data-driven, and repeatable solution, based on ISO standards, and has demonstrated quantifiable value to clients. 

The IDC Metri AVM solutions team began the process of breaking down the product development process and product. 

Measures of value creation occurred across multiple sprints. Enhanced function point analysis, based on ISO standards, is to assess progress and efficiency rather than Story Points. Source code analysis assed code quality and security, to reduce refactoring and the number of CVEs. This analysis not only identified the gaps, but provided specific, tactical recommendations on how to address them. Further, the benchmarking team gauged their performance to understand how competitive their team was versus market peers. 

After six weeks of work, management and team received dashboards and analysis that clearly identified gaps and remediation. 

The engineering dashboards clearly identified poor code and CVEs that allowed the development team to better and more rapidly address quality issues. As they adopted the guidance from the engineering dashboard, the development team found their practices improving. Quality and performance improved due to lower testing efforts attributed to enhanced coding practices. Also, improving practices and identifying better practices reduced team stress, and enabled newly onboarded team members to more rapidly become productive. 

The benchmarking team let management identify areas to invest and realize team improvement. The management dashboard provided managers the opportunity to check key project factors, notice and correct negative performance trends.

Based on the provided forecasting assessment, management realized the development team wasn’t ready for the required business delivery milestones. Even if they performed to peers. So, management made the decision to augment the team with third-party staff to ensure success.

Both the development team and management needed the same thing – a common, agreed-upon definition of “reality”. They started off with two different frames of reference. For the development team, their frame was backlog, sprints, velocity and Story Points. For management it was cost, time, quality and milestones. The two groups were speaking different languages and couldn’t reconcile. By bringing in AVM, management and the development team now had one frame of reference everyone could agree with. They had a complete analysis that brought together technical and business needs.

The team had one language from which they could understand development health, address gaps and deliver agile value.

Interested in learning more about IDC Metri? Let’s schedule an appointment for an introductory meeting.