<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>BMC Software | Blogs</title>
	<atom:link href="https://blogs.bmc.com/feed/" rel="self" type="application/rss+xml" />
	<link>https://s7280.pcdn.co</link>
	<description></description>
	<lastBuildDate>Fri, 10 Apr 2026 11:04:19 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	

<image>
	<url>https://s7280.pcdn.co/wp-content/uploads/2016/04/bmc_favicon-300x300-36x36.png</url>
	<title>BMC Software | Blogs</title>
	<link>https://s7280.pcdn.co</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>A Mainframe Future Built on AI and Open Integration</title>
		<link>https://s7280.pcdn.co/mainframe-future-ai-open-integration/</link>
		
		<dc:creator><![CDATA[Matt Whitbourne]]></dc:creator>
		<pubDate>Wed, 08 Apr 2026 09:41:14 +0000</pubDate>
				<category><![CDATA[Mainframe Blog]]></category>
		<guid isPermaLink="false">https://blogs.bmc.com/?p=55887</guid>

					<description><![CDATA[<img width="810" height="405" src="https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-1024x512.png" class="attachment-large size-large wp-post-image" alt="" decoding="async" fetchpriority="high" srcset="https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-1024x512.png 1024w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-300x150.png 300w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-768x384.png 768w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-810x405.png 810w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-1140x570.png 1140w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-24x12.png 24w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-36x18.png 36w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-48x24.png 48w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion.png 1400w" sizes="(max-width: 810px) 100vw, 810px" />Artificial intelligence (AI) is evolving rapidly, and so are organizations’ approach to its use. In April 2026, BMC released a statement of direction on the integration of AI into the BMC AMI suite of mainframe solutions. In The Next Evolution in Enterprise AI is Purpose Built, BMC mainframe Senior Vice President and General Manager John […]]]></description>
										<content:encoded><![CDATA[<img width="810" height="405" src="https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-1024x512.png" class="attachment-large size-large wp-post-image" alt="" decoding="async" srcset="https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-1024x512.png 1024w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-300x150.png 300w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-768x384.png 768w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-810x405.png 810w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-1140x570.png 1140w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-24x12.png 24w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-36x18.png 36w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion-48x24.png 48w, https://s7280.pcdn.co/wp-content/uploads/2022/12/tb-digital-data-screen-coming-through-speed-motion.png 1400w" sizes="(max-width: 810px) 100vw, 810px" /><p>Artificial intelligence (AI) is evolving rapidly, and so are organizations’ approach to its use. In April 2026, BMC released a statement of direction on the integration of AI into the <a href="/it-solutions/bmc-ami-automated-mainframe-intelligence.html">BMC AMI</a> suite of mainframe solutions. In <a href="/blogs/purpose-built-agentic-ai-mainframe-statement-of-direction">The Next Evolution in Enterprise AI is Purpose Built</a>, BMC mainframe Senior Vice President and General Manager John McKenny discusses the shift from generative AI (GenAI) to agentic AI and our intent to embed agentic AI across the BMC AMI portfolio.</p>
<p>While the statement of direction explains how we’re moving into the future with agentic AI workflows, over the past two years we have developed purpose-built AI for the mainframe, offering in-context expertise through <a href="/documents/solution-briefs/accelerate-mainframe-transformation-bmc-ami-assistant.html">BMC AMI Assistant</a>. With Knowledge Expert Chat, practitioners can harness the power of AI to find the answers they need when and where they need them, increasing the quality and efficiency of their work. Each quarter, we have added to what mainframe professionals can accomplish while increasing integration of BMC AMI Assistant across the BMC AMI portfolio.</p>
<p>The April 2026 release of enhancements to the BMC AMI portfolio focuses on this advancement of mainframe GenAI as well as the open integration of the platform with the broader enterprise IT ecosystem.</p>
<h2>GenAI powered by organizational intelligence</h2>
<p>Even as we increase the use of agent-driven automation on the mainframe, the efforts to improve GenAI assistance continue, including this quarter’s General Availability of Knowledge Hub, which surfaces institutional knowledge from across the organization to offer contextual information and answers at the moment of decision through <a href="https://youtu.be/_opCbtYli-8?si=YfYkzFh7SaaaXQ9V">BMC AMI Assistant Knowledge Expert Chat</a>.</p>
<p>Knowledge Hub works behind the scenes, ingesting provided institutional knowledge from past issue resolutions, operational insights, and information from runbooks, tickets, and shared files, then combining it with BMC mainframe expertise to create Knowledge Expert Chat responses that are contextually aware and tailored to user’s mainframe environment. These responses are available directly in workflows using existing BMC tools, enabling quicker decisions made with expert-level confidence, regardless of the user’s experience level.</p>
<h2>AI-generated application analysis reports</h2>
<p>Just as Knowledge Hub captures institutional knowledge to inform assistance provided by Knowledge Expert Chat, a new feature in <a href="/it-solutions/bmc-ami-zadviser.html">BMC AMI zAdviser Enterprise</a> adds hard-won institutional knowledge to zAdviser’s collection of development tool and toolchain data to help development managers and their developers understand the applications they are modifying.</p>
<p>BMC AMI zAdviser Enterprise gathers BMC AMI tool usage data and DevOps metrics to give development managers a clear picture of development effectiveness. With new application analysis reports, they now get a single AI-generated view of how their applications work, where the risk is, and where their team&#8217;s time is going, accelerating modernization decisions and cutting weeks off of developer onboarding.</p>
<p>BMC AMI zAdviser Enterprise’s Application Analysis turns tribal knowledge into organizational knowledge before it walks out the door. The narrative assessments capture the knowledge of experienced developers, sharing that expertise with development managers and developers, creating more efficient application review, planning, and optimization.</p>
<p>Beyond a greater understanding of individual programs, these application analysis reports provide a clear picture of which programs are attracting a disproportionate share of developer attention because of failures and modifications. By correlating failure history with code complexity and maintenance patterns, they enable development teams to target problem applications with proactive remediation, improving system resilience and development efficiency.</p>
<h2>Standardizing enterprise security</h2>
<p>Manual management of security certificates decreases efficiency and weakens system security. For some time, security teams have been able to automate certificate management on distributed systems using third-party tools.BMC revolutionized mainframe security management with <a href="https://soundcloud.com/modernmainframe/strengthening-mainframe-zero-trust-security-with-automated-certificate-management">BMC AMI Enterprise Connector for Venafi</a>. This April, we introduce BMC AMI Certificate Manager, a new product within <a href="/it-solutions/bmc-ami-mainframe-security.html">BMC AMI Security</a> that gives customers more choice and flexibility in integrating enterprise security solutions with the mainframe.</p>
<p>Designed to integrate with IBM Z<sup>®</sup> enterprise security management (ESM) environments, including RACF<sup>®</sup>, ACF2<sup>™</sup>, and Top Secret<sup>®</sup> for z/OS to automate the full certificate lifecycle without the need for infrastructure changes, BMC AMI Certificate Manager extends leading enterprise certificate management platforms to the mainframe. This gives CISOs and security teams the ability to use the same vendor solutions for distributed and mainframe certificate management, further integrating the mainframe with—and simplifying—enterprise security efforts.</p>
<p>BMC AMI Certificate Manager currently integrates with Venafi<sup>®</sup> and Keyfactor<sup>®</sup>, with further integrations to follow in future releases. With this unified integration layer between enterprise certificate management tools and IBM Z<sup>®</sup> ESMs, organizations can automate certificate issuance, renewal, and enforcement with one BMC solution while supporting multiple vendors.</p>
<h2>Breaking down silos with new technology</h2>
<p>This quarter’s enhancements further BMC’s commitment to optimizing what is possible on the mainframe through open integration of the platform and continuous improvements to AI capabilities. With the addition of organization-specific knowledge to AI engines, Knowledge Hub empowers mainframe professionals to make the right decisions as they do their jobs, regardless of their experience and skill levels. Application analysis reports in BMC AMI zAdviser Enterprise combine development and application performance data with system telemetry to provide a clear picture of how applications are performing and where attention and efforts should be focused. And the new BMC AMI Certificate Manager enables security teams to employ certificate management policies across platforms without the need for separate tooling.</p>
<p>Each of these enhancements improves the productivity and efficiency of mainframe teams, a goal BMC is committed to pursuing with each of our quarterly releases. Make your mainframe the engine of faster, better, and smarter answers. When you BMC First.</p>
<p>These are just a few of the innovations included in the April 2026 release of BMC AMI features. To learn more about everything included in the release, visit the <a href="/it-solutions/bmc-ami-latest-release.html">What’s New in Mainframe Solutions webpage</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The Next Evolution in Enterprise AI is Purpose-Built</title>
		<link>https://blogs.bmc.com/purpose-built-agentic-ai-mainframe-statement-of-direction/</link>
		
		<dc:creator><![CDATA[John McKenny]]></dc:creator>
		<pubDate>Wed, 08 Apr 2026 09:39:55 +0000</pubDate>
				<category><![CDATA[Mainframe Blog]]></category>
		<guid isPermaLink="false">https://blogs.bmc.com/?p=55892</guid>

					<description><![CDATA[<img width="810" height="405" src="https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-1024x512.jpg.optimal.jpg" class="attachment-large size-large wp-post-image" alt="" decoding="async" srcset="https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-1024x512.jpg.optimal.jpg 1024w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-300x150.jpg.optimal.jpg 300w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-768x384.jpg.optimal.jpg 768w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-810x405.jpg.optimal.jpg 810w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-1140x570.jpg.optimal.jpg 1140w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-24x12.jpg.optimal.jpg 24w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-36x18.jpg.optimal.jpg 36w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-48x24.jpg.optimal.jpg 48w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain.jpg.optimal.jpg 1400w" sizes="(max-width: 810px) 100vw, 810px" />Mainframe organizations continue to run some of the most critical systems in the global economy while managing increasing complexity and rising expectations for innovation. Artificial intelligence (AI) provides a practical way to scale expertise, reduce operational friction, and help teams move work forward with greater confidence. BMC believes organizations will increasingly leverage AI to broaden […]]]></description>
										<content:encoded><![CDATA[<img width="810" height="405" src="https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-1024x512.jpg.optimal.jpg" class="attachment-large size-large wp-post-image" alt="" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-1024x512.jpg.optimal.jpg 1024w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-300x150.jpg.optimal.jpg 300w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-768x384.jpg.optimal.jpg 768w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-810x405.jpg.optimal.jpg 810w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-1140x570.jpg.optimal.jpg 1140w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-24x12.jpg.optimal.jpg 24w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-36x18.jpg.optimal.jpg 36w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain-48x24.jpg.optimal.jpg 48w, https://s7280.pcdn.co/wp-content/uploads/2026/04/brain.jpg.optimal.jpg 1400w" sizes="auto, (max-width: 810px) 100vw, 810px" /><p>Mainframe organizations continue to run some of the most critical systems in the global economy while managing increasing complexity and rising expectations for innovation. Artificial intelligence (AI) provides a practical way to scale expertise, reduce operational friction, and help teams move work forward with greater confidence.</p>
<p>BMC believes organizations will increasingly leverage AI to broaden access to mainframe capabilities and integrate these systems more fully into their broader enterprise IT ecosystems. In many cases, AI will also act as a catalyst for modernization in place—helping organizations evolve applications, workflows, and operational practices while continuing to run critical workloads on the platform that already powers their business.</p>
<h2>The Direction: From AI assistance to coordinated intelligence</h2>
<p>Enterprise AI is moving beyond generating insights and explanations. Organizations now expect intelligence that can safely participate in execution across critical systems. <strong>Our direction is clear: </strong><a href="/it-solutions/bmc-ami-automated-mainframe-intelligence.html">BMC AMI</a> solutions are evolving into intelligent participants in a governed ecosystem of AI agents that collaborate autonomously across development, operations, data, and security workflows to accelerate innovation on the mainframe.</p>
<p>Supporting this direction requires enterprise AI on the mainframe to operate across three coordinated layers:</p>
<p><strong>Intelligence layer</strong> — Domain knowledge and reasoning grounded in mainframe expertise, operational telemetry, and institutional knowledge.</p>
<p><strong>Coordination layer</strong> — Orchestrated AI agents collaborating across workflows to connect understanding with action.</p>
<p><strong>Governance layer</strong> — Policy-aware controls ensuring AI-driven actions remain secure, transparent, auditable, and subject to human oversight.</p>
<p>Together, these layers enable AI to extend beyond analysis and participate responsibly in execution. This foundation now enables the next phase of our direction: extending AI beyond knowledge and insight toward coordinated, governed execution across the BMC AMI platform.</p>
<h2>AI in production: Trust earned, not claimed</h2>
<p>In July 2024, we published a <a href="/blogs/bmc-ami-platform-statement-direction-ai-cloud/">Statement of Direction</a> focused on <a href="https://techstrong.tv/videos/interviews/infusing-intelligence-ai-as-a-partner-genai-and-the-modern-mainframe-bmc-software" target="_blank" rel="noopener">infusing intelligence</a> directly into BMC AMI product experiences. Our objective was straightforward: reduce complexity, close knowledge gaps, and make it easier for teams to work confidently on the mainframe.</p>
<p>That direction was grounded in three core principles:</p>
<ul>
<li><strong>Consistent and contextual intelligence</strong>, with a common AI foundation infused across the AMI portfolio</li>
<li><strong>LLM freedom</strong>, allowing organizations to choose the models that align with their enterprise AI strategies</li>
<li><strong>In-the-moment intelligence</strong>, embedding AI directly into workflows to surface answers and insights where work happens</li>
</ul>
<p>Since then, we have delivered.</p>
<p><a href="/it-solutions/mainframe-ai.html">BMC AMI Assistant</a> was introduced beginning with <a href="https://techstrong.tv/videos/interviews/generative-ai-and-mainframe-evolution-with-bmcs-anthony-distauro" target="_blank" rel="noopener">GenAI-powered code explanation</a> and expanded across consecutive quarterly releases into a cohesive intelligence layer embedded throughout the BMC AMI portfolio. Using <a href="https://techstrong.tv/videos/interviews/laying-the-groundwork-ai-as-an-advisor-bmc-software" target="_blank" rel="noopener">AI as an advisor</a>, developers gained faster understanding of unfamiliar applications, while operators gained explainable insight into system and operational issues. Investigation that once required manual analysis increasingly includes guided insight and next steps directly inside the workflow.</p>
<p>The addition of <a href="https://www.linkedin.com/pulse/answers-seconds-how-ai-closing-mainframe-skills-gap-eric-odell-x5mif" target="_blank" rel="noopener">Knowledge Expert Chat and Knowledge Hub</a> further expanded this intelligence layer. Together, these capabilities combine curated product intelligence, decades of BMC mainframe expertise, and each organization’s own institutional knowledge, making trusted guidance available at the moment of decision. <strong>This approach helps less-experienced staff work with greater confidence while allowing senior experts to focus on higher-value work, turning institutional knowledge into a shared operational asset.</strong> Watch this <a href="https://www.youtube.com/watch?v=xzfFcxX9rI4" target="_blank" rel="noopener">short video</a> for a demo of how Knowledge Expert Chat provides answers at the moment of need.</p>
<p>Instead of searching across documentation, runbooks, and tickets, teams can surface answers through natural-language questions using a <a href="https://www.youtube.com/watch?v=SIcoqN9eKG4" target="_blank" rel="noopener">knowledge expert</a> chat inside their workflow. Early usage shows that this approach not only accelerates troubleshooting and decision-making but also helps users discover capabilities and documentation they did not previously know existed.</p>
<p>Trust in AI is earned through real outcomes. This intelligence foundation now supports the next phase of our direction.</p>
<h2>Customers are redefining what AI must do</h2>
<p>Through advisory boards, design programs, beta participation, and enterprise <a href="https://hyperframeresearch.com/wp-content/uploads/2026/02/HFR_BMC_-Beyond-GenAI_Draft_v2-021626.pdf" target="_blank" rel="noopener">AI readiness</a> discussions, a clear message has emerged: customers want AI that does more than explain what happened. They want orchestrated intelligence that helps move work forward.</p>
<p>They want to reduce repetitive analysis, move faster from insight to action, and make it easier for teams to build expertise and focus on higher-value priorities. They want to realize the full promise of <a href="/it-solutions/bmc-ami-automated-mainframe-intelligence.html">BMC AMI as automated mainframe intelligence</a>.</p>
<p>Generative AI laid the foundation by helping teams understand, explain, and accelerate work. Agentic AI builds on that foundation by enabling intelligence to plan, decide, and safely carry work forward.</p>
<p>To meet these expectations at enterprise scale, our new Statement of Direction is grounded in five core principles:</p>
<ul>
<li><strong>Establish the core first:</strong> Build an orchestrated agentic intelligence foundation.</li>
<li><strong>Specialized agents:</strong> Reflect domain expertise with focused, autonomous actions.</li>
<li><strong>Human-in-the-loop:</strong> Governed, transparent, and observable autonomy.</li>
<li><strong>Open ecosystem:</strong> Open standards-based, composable, and extensible.</li>
<li><strong>Outcome-driven adoption:</strong> Deliver measurable value early and expand over time.</li>
</ul>
<p>The next phase extends AI beyond insight toward trusted participation in execution, within enterprise guardrails.</p>
<p>And that shift defines our next statement of direction.</p>
<h2>From AI assistance to orchestrated intelligence</h2>
<p>Building on the foundation of the AI-driven knowledge and guidance we have delivered with BMC AMI Assistant, our next statement of direction is clear: extending that foundation toward proactive, coordinated intelligence that can safely execute actions within defined governance policies across the BMC AMI portfolio.</p>
<p>This evolution introduces coordinated AI agents with specialized, policy-aware capabilities operating across development, operations, data, and security. Each agent is domain-specific, working together within enterprise guardrails to deliver governed, collaborative outcomes.</p>
<p>In this next phase, AI agents for operations, development, data, and security will operate as orchestrated participants across domains—supporting actions such as system and performance diagnostics, development workflows, security validation, and operational recovery. They will learn across past incidents and participate in execution with transparency, verifiability, enterprise governance, and human oversight built in by design.</p>
<p>We will expand beyond explanation and recommendation to enable validated action inside workflows. Agentic workflows will move from detection to investigation to explanation to resolution within a single, governed AI experience.</p>
<p>AI will not replace expertise. Human validation remains essential. Principal engineers and system programmers will continue to define policy, validate outcomes, and shape how intelligence operates. AI scales expertise rather than removing it.</p>
<p>The result is faster problem identification, more contextual decision-making, and reduced operational friction—while ensuring the mainframe participates fully in broader enterprise AI strategies. The mainframe will operate as a connected, policy-aligned participant in the enterprise AI ecosystem across hybrid environments, maintaining the security, reliability, and trust on which organizations depend.</p>
<h2>Delivering agentic workflows across the BMC AMI portfolio</h2>
<p>Over the coming releases, this direction will take shape through <strong>agentic workflows that coordinate knowledge, reasoning, and action across BMC AMI solutions</strong>. Initial workflows could include agentic AIOps incident resolution, agentic diagnostics, AI-assisted application insights, development troubleshooting, and test-case generation.</p>
<p>These workflows build on the <strong>intelligence layer</strong> already established across the BMC AMI portfolio, combining development and operations telemetry, mainframe domain expertise, and organizational knowledge to provide the context required for responsible execution.</p>
<p>From there, <a href="https://techstrong.tv/videos/interviews/acting-with-confidence-ai-as-an-agent-of-change-bmc-software" target="_blank" rel="noopener">coordinated AI agents operate across solutions,</a> aligning development, operations, data, and security activities that were previously siloed. Instead of isolated AI features, intelligence becomes part of the workflow itself, helping teams connect system understanding with the next operational step.</p>
<p>Every action remains governed and transparent. <strong>Human validation, enterprise policy, and operational guardrails remain central</strong>, ensuring AI participation strengthens reliability rather than introducing risk.</p>
<p>The objective is clear: translate insight into trusted action while preserving the control and discipline enterprise systems require.</p>
<h2>Establishing the foundation for governed AI execution</h2>
<p>As AI becomes more operational, governance and trust become increasingly essential. Organizations expect AI not only to inform decisions, but to operate safely, predictably, and transparently within enterprise policy boundaries.</p>
<p>Across the industry, fragmented approaches are already emerging: isolated AI integrations, independently managed execution layers, and disconnected tool endpoints. While intended to accelerate innovation, these models often introduce new complexity, inconsistent policy enforcement, and operational sprawl.</p>
<p>Enterprise mainframe environments cannot afford that fragmentation, especially as AI Agent orchestration becomes a core capability for coordinating multi-step agentic workflows. This can result in agents giving conflicting recommendations and teams lose confidence and revert to manual work. Without a unified and governed approach, orchestration itself becomes fragmented, leading to unpredictable behavior, loss of insight, and reduced trust in AI-driven operations.</p>
<p>As part of our next phase, our direction includes establishing an <strong>MCP Gateway</strong> as a shared, governed access layer across the BMC AMI portfolio. Rather than creating multiple independently governed AI entry points, the MCP Gateway will provide a centralized, secure, and policy-aware interface through which AI agents interact with BMC AMI solutions. Every AI-driven action will be visible, governed, and aligned to enterprise policy with consistent controls across the platform. This enables AI agents not only to access BMC AMI systems, but to safely execute actions across them within defined policy boundaries.</p>
<p>Supporting this architecture, we will introduce an <strong>Agent Gateway</strong> to facilitate how AI agents communicate and collaborate. Instead of agents interacting independently, interactions will flow through the Agent Gateway—where the agent interactions are visible—ensuring governance, auditability, logging, and policy enforcement across agentic workflows.</p>
<p>Together, the MCP Gateway and Agent Gateway extend the BMC AMI Platform into a governed AI foundation that enables coordinated intelligence and trusted execution across the portfolio (see Figure 1). The BMC AMI Platform serves as the enterprise foundation for this direction—providing a unified layer of core capabilities and services that connects intelligence, orchestration, and governance across the BMC AMI portfolio. It brings together innovations such as BMC AMI Assistant, the Agent Gateway, and the MCP Gateway to simplify mainframe transformation, accelerate innovation, and enable AI to operate consistently and securely at scale.</p>
<p>Industry standards define how AI systems connect to tools and agents. Our direction focuses on how those connections are governed, secured, and operationalized at enterprise scale—without introducing fragmentation or operational risk.</p>
<p>This is how we intend to evolve agentic AI into an <em>enterprise-grade foundation</em> across the BMC AMI portfolio (see Figure 1).</p>
<p><img loading="lazy" decoding="async" class="alignnone wp-image-55893 size-full" src="https://s7280.pcdn.co/wp-content/uploads/2026/04/high-level-overview.png" alt="" width="624" height="352" srcset="https://s7280.pcdn.co/wp-content/uploads/2026/04/high-level-overview.png 624w, https://s7280.pcdn.co/wp-content/uploads/2026/04/high-level-overview-300x169.png 300w, https://s7280.pcdn.co/wp-content/uploads/2026/04/high-level-overview-24x14.png 24w, https://s7280.pcdn.co/wp-content/uploads/2026/04/high-level-overview-36x20.png 36w, https://s7280.pcdn.co/wp-content/uploads/2026/04/high-level-overview-48x27.png 48w" sizes="auto, (max-width: 624px) 100vw, 624px" /></p>
<p><em>Figure 1: High-level overview of BMC AMI Platform’s agentic architecture supporting enterprise-grade AI.</em></p>
<h2>What you can expect — and how to shape what comes next</h2>
<p>Over the coming releases, you will experience a meaningful shift in how AI participates in the BMC AMI environment. Intelligence will extend beyond explanation and guidance to support the safe execution of operational tasks. Capabilities that once appeared as isolated features will increasingly operate as coordinated agentic workflows across the BMC AMI portfolio, connecting development, operations, data, and security activities.</p>
<p>AI will participate responsibly within enterprise guardrails, operating under the policies, validation models, and controls defined by the teams who run these systems. This direction represents more than a single capability release, it establishes the foundation for a new way of working with BMC AMI solutions.</p>
<p>We invite customers to help us shape this next phase. You can engage with us in several ways: join our <a href="/info/customer-design-partnership.html">Customer Design Partner program</a> to help refine and validate the most impactful agentic workflows, participate in early access initiatives focused on execution-oriented capabilities, and work with us to shape the governance and policy models that will guide enterprise-scale AI execution.</p>
<h2>Moving forward with confidence</h2>
<p>We delivered on our commitments. We established a foundation of trust. And we are now stepping into the next phase with clarity.</p>
<p>The future of AI on the mainframe centers on purpose-built enterprise intelligence, where your choice of AI models, your institutional knowledge and operational practices, your people, and your chosen platforms work together to drive intelligent execution across the enterprise. In this model, the mainframe is not an isolated environment. It is a fully governed participant in enterprise AI strategies, capable of supporting intelligent execution across hybrid systems.</p>
<p>Organizations that embrace this direction will not simply modernize their mainframe environments. They will unlock new ways to operate complex workloads with greater visibility, control, and intelligence.</p>
<p>And this is only the beginning. Before you run AI on your most essential platform, BMC First.</p>
<p>Are you truly ready to move from AI insight to AI-driven execution on the mainframe?<br />
Assess your organization’s AI readiness and maturity level—register for an <a href="/forms/building-ai-ready-mainframe-foundations.html"><strong>AI Readiness Discovery Workshop</strong></a> to take the next step.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Service Orchestration Solves What Scheduling Cannot</title>
		<link>https://blogs.bmc.com/service-orchestration-not-job-scheduling/</link>
		
		<dc:creator><![CDATA[Basil Faruqui]]></dc:creator>
		<pubDate>Fri, 03 Apr 2026 13:39:29 +0000</pubDate>
				<category><![CDATA[Workload Automation Blog]]></category>
		<guid isPermaLink="false">https://blogs.bmc.com/?p=55885</guid>

					<description><![CDATA[<img width="700" height="400" src="https://s7280.pcdn.co/wp-content/uploads/2019/07/All-DB2-DBAs-497452519-700x400.png" class="attachment-large size-large wp-post-image" alt="" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2019/07/All-DB2-DBAs-497452519-700x400.png 700w, https://s7280.pcdn.co/wp-content/uploads/2019/07/All-DB2-DBAs-497452519-700x400-300x171.png 300w, https://s7280.pcdn.co/wp-content/uploads/2019/07/All-DB2-DBAs-497452519-700x400-24x14.png 24w, https://s7280.pcdn.co/wp-content/uploads/2019/07/All-DB2-DBAs-497452519-700x400-36x21.png 36w, https://s7280.pcdn.co/wp-content/uploads/2019/07/All-DB2-DBAs-497452519-700x400-48x27.png 48w" sizes="auto, (max-width: 700px) 100vw, 700px" />Most organizations don’t start by thinking about service orchestration. They start with symptoms: a morning dashboard showing incomplete data, an SLA missed because an upstream process ran late, or a cross-team Slack thread at 7 AM trying to figure out what broke and why.  By the time the word “orchestration” enters the conversation, the workarounds […]]]></description>
										<content:encoded><![CDATA[<img width="700" height="400" src="https://s7280.pcdn.co/wp-content/uploads/2019/07/All-DB2-DBAs-497452519-700x400.png" class="attachment-large size-large wp-post-image" alt="" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2019/07/All-DB2-DBAs-497452519-700x400.png 700w, https://s7280.pcdn.co/wp-content/uploads/2019/07/All-DB2-DBAs-497452519-700x400-300x171.png 300w, https://s7280.pcdn.co/wp-content/uploads/2019/07/All-DB2-DBAs-497452519-700x400-24x14.png 24w, https://s7280.pcdn.co/wp-content/uploads/2019/07/All-DB2-DBAs-497452519-700x400-36x21.png 36w, https://s7280.pcdn.co/wp-content/uploads/2019/07/All-DB2-DBAs-497452519-700x400-48x27.png 48w" sizes="auto, (max-width: 700px) 100vw, 700px" /><p><span data-contrast="auto">Most organizations don’t start by thinking about service orchestration. They start with symptoms: a morning dashboard showing incomplete data, an SLA missed because an upstream process ran late, or a cross-team Slack thread at 7 AM trying to figure out what broke and why.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">By the time the word “orchestration” enters the conversation, the workarounds have usually been in place for months. And that’s exactly the problem—those workarounds are invisible because they look like “process.”</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">The real distinction isn’t between working and broken systems. It’s between environments that are quietly compensating for a lack of coordination and those built on a foundation that can actually support end-to-end execution. That’s where the difference between scheduling and orchestration begins—and why it matters.</span></p>
<h2><span class="TextRun SCXW121351974 BCX8" lang="EN" xml:lang="EN" data-contrast="auto"><span class="NormalTextRun SCXW121351974 BCX8">Where Job Scheduling Hits Its Ceiling</span></span></h2>
<p><span data-contrast="auto">Technically there&#8217;s nothing wrong with the built-in automation tools that come with modern applications. ERP systems, CRMs, data warehouses, and large database platforms all ship with native scheduling capability. For environments that aren&#8217;t particularly complex, those tools do exactly what they&#8217;re supposed to do: automate jobs within the boundaries of the application that owns them.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">The problem starts at the boundary.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">When a business outcome requires work that transcends multiple applications &#8211; supply chain operations, financial close, ML pipeline execution &#8211; there is no common coordination layer to manage it. Each application scheduler knows its own world and nothing else. The result is a set of individually automated workflows that have no shared understanding of each other&#8217;s state, progress, or failure.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">This is the core architectural gap that service orchestration addresses. A control plane above existing schedulers manages end-to-end execution with business objectives as the organizing principle.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<h2><span class="TextRun SCXW170207398 BCX8" lang="EN" xml:lang="EN" data-contrast="auto"><span class="NormalTextRun SCXW170207398 BCX8" data-ccp-parastyle="heading 2">What Cross-System Coordination Actually Looks Like in Practice</span></span></h2>
<p><span data-contrast="auto">A customer submits an application through a web portal. From there, the process fans out across a stack of systems that have very little in common: data validation runs through an API service in Kubernetes containers; identity verification calls out to a third-party SaaS provider; risk scoring runs on a model deployed in the cloud; the actual account gets created in a core banking system that lives on-premises; the CRM updates the customer profile; compliance documentation gets generated and archived.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">That&#8217;s six distinct environments &#8211; cloud, on-premises, containerized infrastructure, SaaS &#8211; all participating in a single business transaction. And that&#8217;s the </span><b><span data-contrast="auto">simplified</span></b><span data-contrast="auto"> version.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Now add the business constraint that makes this genuinely difficult to manage: the customer expects the whole process to complete within one hour of submitting their application. To be clear, that&#8217;s a business commitment &#8211; and it applies to the entire chain, not to any individual step.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">No application scheduler was built to coordinate that topology under that deadline. Each one was designed to manage its own job queue, not to understand where it sits in a larger business process or what happens downstream if it falls behind.</span></p>
<h2><span class="TextRun SCXW48924073 BCX8" lang="EN" xml:lang="EN" data-contrast="auto"><span class="NormalTextRun SCXW48924073 BCX8" data-ccp-parastyle="heading 2">How the Problem Reveals Itself Operationally</span></span></h2>
<p><span data-contrast="auto">Organizations don&#8217;t usually discover this gap through architectural reviews. They discover it through patterns of operational friction that gradually become normalized.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">The most telling early signal is what might be called the time buffer contract. When data teams and application teams operate without a common coordination layer, they start negotiating informal timing agreements. The data from the application layer usually lands by 4 PM, so the ETL pipelines are configured to kick off at 5, with an hour of buffer built in as insurance.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">That buffer looks like responsible planning, but it masks missing dependency management. It fails in both directions &#8211; if the upstream data is late, the downstream pipeline starts anyway and processes whatever is there; if the data arrives early, the pipeline sits idle </span><b><span data-contrast="auto">waiting for a clock to tick</span></b><span data-contrast="auto">.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">The second pattern is more disruptive. A data pipeline ingests records from a CRM and an ERP expecting a complete dataset &#8211; a million records, in a typical case. An upstream workflow encountered a problem, so now the data arrives incomplete. The pipeline has no way to know this, so it runs to completion and updates business dashboards with partial data. No one finds out until a business user notices something looks wrong.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">What follows is a cross-team reconstruction effort: Slack messages, conference calls, manual tracing of execution steps across multiple tools to figure out where in the chain the problem originated. This kind of incident response &#8211; teams scrambling across disconnected systems to piece together what happened &#8211; is one of the clearest indicators that an organization has outgrown its current approach to workflow coordination.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">The absence of process lineage across the full stack &#8211; from the business application layer down through the data layer &#8211; is what makes these incidents so difficult to resolve. As Basil Faruqui, Sr. Director of Product Marketing at BMC Software, puts it: &#8220;You cannot manage what you can&#8217;t see.&#8221; And you can&#8217;t see what no single tool was designed to show you.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<h2><span class="TextRun SCXW93466376 BCX8" lang="EN" xml:lang="EN" data-contrast="auto"><span class="NormalTextRun SCXW93466376 BCX8" data-ccp-parastyle="heading 2">What Orchestration Requires in the Modern Era</span></span><span class="EOP SCXW93466376 BCX8" data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;201341983&quot;:0,&quot;335559738&quot;:360,&quot;335559739&quot;:120,&quot;335559740&quot;:240}"> </span></h2>
<p><span data-contrast="auto">Listing orchestration capabilities as a feature set understates what&#8217;s actually required. The architecture needs to be a direct response to the failure modes that siloed scheduling creates.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">The first requirement is broad environmental support. A control plane that only works in cloud environments, or only connects to certain application types, is just another silo with a better UI.</span></p>
<p><span data-contrast="auto">Orchestrating a business outcome end-to-end is complex and needs to span SaaS applications, multi-cloud infrastructure, on-premises data centers, and even mainframe systems in some cases. For many industries like banking, financial services and insurance mainframe support isn&#8217;t optional.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">The second requirement is end-to-end visibility with process lineage. Knowing that a job completed isn&#8217;t the same as knowing whether the business service it belongs to is on track. Effective visibility means being able to trace the state of a workflow across every system it touches, in real time.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">The third requirement is SLA management that carries business context. A notification that says a process is running late is less useful than a notification that says a supply chain workflow is running late and will affect five downstream business services if it doesn&#8217;t recover within the next 20 minutes. The former tells you something is wrong. The latter tells you </span><b><span data-contrast="auto">what</span></b><span data-contrast="auto"> to do about it.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Self-healing is the fourth requirement, and it&#8217;s where organizational knowledge becomes operational policy. When a workflow step fails because of a transient network issue, someone investigates and decides to rerun the step. That resolution should be capturable as a policy &#8211; and increasingly, AI-enabled capabilities can identify that the same failure pattern has occurred before, surface the historical remediation, and automate the response with human approval where appropriate. The goal is to build institutional knowledge into the platform rather than relying on it to live only in the heads of the people who were on shift when the incident happened.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Finally, CI/CD integration matters more than it&#8217;s often given credit for. Orchestration is, at its core, </span><b><span data-contrast="auto">a team sport</span></b><span data-contrast="auto">. Business users, application teams, and operations teams all have a stake in how workflows are defined, tested, and deployed. Treating orchestration workflows as code &#8211; with version control, automated testing, and deployment pipelines consistent with modern DevOps practice &#8211; isn&#8217;t a nice-to-have. It&#8217;s what makes the platform manageable at all scales.</span></p>
<h2><span class="TextRun SCXW75699846 BCX8" lang="EN" xml:lang="EN" data-contrast="auto"><span class="NormalTextRun SCXW75699846 BCX8" data-ccp-parastyle="heading 2">AI Workloads Break the Pipeline Model</span></span><span class="EOP SCXW75699846 BCX8" data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;201341983&quot;:0,&quot;335559738&quot;:360,&quot;335559739&quot;:120,&quot;335559740&quot;:240}"> </span></h2>
<p><span data-contrast="auto">One of the more common misconceptions in this space is that AI and ML workloads can be handled by extending existing data pipeline tools. The assumption is that model training is just a compute-intensive batch job, and inference is just another API call. Both assumptions fall short.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Consider a customer churn detection model. The training data has to be sourced from systems that aren&#8217;t data systems at all &#8211; ERP, CRM, custom applications, potentially social media feeds. Getting that data into a staging environment already requires coordination across the application layer. Then the model training pipelines need to run. So far, that&#8217;s two layers of coordination before any inference happens.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Once the model is in production and identifies customers at risk, the business response &#8211; sending a promotional offer, for example &#8211; doesn&#8217;t happen inside the model. It happens in the CRM and ERP layer where customer profiles and promotional workflows live. The model is one step in a business process that starts with applications and ends with a business action.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Confining AI workloads to a data pipeline tool means the platform can manage the middle of that chain but not the ends. The data sourcing and the downstream business response are both outside its scope. Siloing the AI workload in a single tool leaves you with a system that can&#8217;t map or control the process from source to business action.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">This matters even more as agent-based architectures become more common. Organizations are already deploying agents that take on discrete tasks within larger processes. A workflow like order-to-cash will not be handed over to agents in its entirety anytime soon &#8211; but agents are increasingly handling specific components of it. That means an orchestration platform needs to be able to invoke agents, pass them the right context, enforce guardrails, and manage their execution according to the same SLAs that govern the rest of the workflow. For organizations with active agent deployments, agent orchestration is already a present consideration.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<h2><span class="TextRun SCXW93885898 BCX8" lang="EN" xml:lang="EN" data-contrast="auto"><span class="NormalTextRun SCXW93885898 BCX8" data-ccp-parastyle="heading 2">Rethinking the Question</span></span></h2>
<p><span data-contrast="auto">When organizations run into workflow coordination problems, the instinct is to ask whether they need better automation. More often, what they actually need is orchestration.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Better automation improves individual jobs. Orchestration manages the relationships between them. For workflows that stay within the boundaries of a single application, better automation is probably the right answer. For workflows that span systems, teams, and environments to deliver a business outcome with a defined deadline, automation alone can&#8217;t provide what&#8217;s needed.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">The symptoms are identifiable. Time buffers masquerading as process discipline. Partial data incidents that require multi-team investigation to trace. SLA breaches that can&#8217;t be attributed to a specific failure point because no single tool has visibility across the full chain. ​​None of these are signs that automation needs tuning. They point to a missing control plane.</span><span data-ccp-props="{&quot;201341983&quot;:0,&quot;335559738&quot;:240,&quot;335559739&quot;:240,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">The distinction matters because building more automation on top of a coordination gap doesn&#8217;t close the gap. It adds more things to coordinate.</span></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>What Is a Data Pipeline? A Complete Guide</title>
		<link>https://blogs.bmc.com/data-pipeline/</link>
		
		<dc:creator><![CDATA[Jonathan Johnson]]></dc:creator>
		<pubDate>Tue, 31 Mar 2026 00:00:35 +0000</pubDate>
				<category><![CDATA[Machine Learning & Big Data Blog]]></category>
		<guid isPermaLink="false">https://www.bmc.com/blogs/?p=17743</guid>

					<description><![CDATA[<img width="700" height="400" src="https://s7280.pcdn.co/wp-content/uploads/2020/06/speeding-car-lights.jpg.optimal.jpg" class="attachment-large size-large wp-post-image" alt="" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2020/06/speeding-car-lights.jpg.optimal.jpg 700w, https://s7280.pcdn.co/wp-content/uploads/2020/06/speeding-car-lights-300x171.jpg.optimal.jpg 300w, https://s7280.pcdn.co/wp-content/uploads/2020/06/speeding-car-lights-24x14.jpg.optimal.jpg 24w, https://s7280.pcdn.co/wp-content/uploads/2020/06/speeding-car-lights-36x21.jpg.optimal.jpg 36w, https://s7280.pcdn.co/wp-content/uploads/2020/06/speeding-car-lights-48x27.jpg.optimal.jpg 48w" sizes="auto, (max-width: 700px) 100vw, 700px" />A data pipeline is a series of automated steps for moving data from one or more sources to a designated destination, often transforming it along the way. Raw, disparate pieces of data enter one end, undergo processes like cleaning, restructuring, and enrichment, and emerge at the other end as usable insights. Data pipelines are the foundation of […]]]></description>
										<content:encoded><![CDATA[<img width="700" height="400" src="https://s7280.pcdn.co/wp-content/uploads/2020/06/speeding-car-lights.jpg.optimal.jpg" class="attachment-large size-large wp-post-image" alt="" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2020/06/speeding-car-lights.jpg.optimal.jpg 700w, https://s7280.pcdn.co/wp-content/uploads/2020/06/speeding-car-lights-300x171.jpg.optimal.jpg 300w, https://s7280.pcdn.co/wp-content/uploads/2020/06/speeding-car-lights-24x14.jpg.optimal.jpg 24w, https://s7280.pcdn.co/wp-content/uploads/2020/06/speeding-car-lights-36x21.jpg.optimal.jpg 36w, https://s7280.pcdn.co/wp-content/uploads/2020/06/speeding-car-lights-48x27.jpg.optimal.jpg 48w" sizes="auto, (max-width: 700px) 100vw, 700px" /><p><span class="TextRun SCXW223972611 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW223972611 BCX0" data-ccp-parastyle="Normal (Web)">A data pipeline is a series of automated steps for moving data from one or more sources to a designated destination, often transforming it along the way. Raw, disparate pieces of data enter one end, undergo processes like cleaning, restructuring, and enrichment, and </span><span class="NormalTextRun SCXW223972611 BCX0" data-ccp-parastyle="Normal (Web)">emerge</span><span class="NormalTextRun SCXW223972611 BCX0" data-ccp-parastyle="Normal (Web)"> at the other end as usable insights. Data pipelines are the foundation of every analytics dashboard, machine learning model, and data-driven operational decision.</span></span><span class="EOP Selected SCXW223972611 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p aria-level="3"><span class="TextRun SCXW161912634 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW161912634 BCX0" data-ccp-parastyle="Normal (Web)">You could think of a data pipeline like an airport baggage system: bags (data) enter the conveyor system, get scanned (validation), sorted (transformation), and routed to the correct flight (destination database). If one belt jams, everything backs up—just like a bottleneck in a data pipeline.</span></span></p>
<h2><span class="TextRun SCXW139477488 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW139477488 BCX0" data-ccp-parastyle="heading 2">What&#8217;s</span><span class="NormalTextRun SCXW139477488 BCX0" data-ccp-parastyle="heading 2"> the difference between a data pipeline and </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW139477488 BCX0" data-ccp-parastyle="heading 2">ETL</span><span class="NormalTextRun SCXW139477488 BCX0" data-ccp-parastyle="heading 2">?</span></span></h2>
<p><span class="TextRun SCXW236809083 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW236809083 BCX0" data-ccp-parastyle="Normal (Web)">A data pipeline and an ETL pipeline are not the same thing, though the terms are often used interchangeably.</span></span></p>
<p><span data-contrast="auto">A data pipeline is an umbrella term for any set of processes that move data from one system to another. This includes simple data ingestion, real-time streaming, batch processing, and complex multi-step workflows.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">An ETL pipeline (Extract, Transform, Load) is a specific type of data pipeline. Its purpose is to extract data from sources, transform it into the right format, and load it into a destination system like a data warehouse or database. All ETL pipelines are data pipelines, but not all data pipelines are ETL pipelines. A pipeline that moves raw data without transforming it, or streams data in real time, is still a data pipeline—but not an ETL pipeline.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3><span class="TextRun SCXW24509487 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW24509487 BCX0" data-ccp-parastyle="heading 3">Is SQL a data pipeline?</span></span></h3>
<p><span data-contrast="auto">No. SQL (Structured Query Language) is a language used to query, manage, and manipulate data in relational databases—it&#8217;s a tool, not a workflow.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">A data pipeline is the full automated process that moves data from one place to another. SQL can be used inside a data pipeline to filter, join, or transform data, but SQL doesn&#8217;t constitute the pipeline itself. If you&#8217;re building a house, SQL is the hammer and saw—essential tools, but not the entire construction project.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h2><span class="TextRun SCXW242537298 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW242537298 BCX0" data-ccp-parastyle="heading 2">Why do data pipelines matter?</span></span></h2>
<p><span data-contrast="auto">Data pipelines are essential because organizations collect data from too many sources—customer interactions, social media, sales transactions, website logs, IoT devices, and internal applications—to manage manually. Without a systematic way to collect, process, and deliver this data, it quickly becomes unmanageable rather than useful.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Data pipelines power the data-driven decisions we encounter every day: personalized recommendations, real-time fraud alerts, and analytics dashboards. For </span><a href="/info/dataops.html"><span data-contrast="none">DataOps</span></a><span data-contrast="auto"> teams specifically, pipelines help ensure the reliability, scalability, and governance that organizations need to use data confidently instead of being overwhelmed by it.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h2><span class="TextRun SCXW111166561 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW111166561 BCX0" data-ccp-parastyle="heading 2">What are the benefits of a well-executed data pipeline?</span></span></h2>
<p><span data-contrast="auto">Data pipelines don&#8217;t just move data—they make it fit for purpose, delivering it where and when it&#8217;s needed. Five benefits go beyond the basic transport function:</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Enabling analytics and business intelligence: Pipelines feed cleaned, structured data into data warehouses and analytical platforms, allowing analysts to discover trends, identify opportunities, and monitor performance.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Fueling machine learning and AI: AI models require large volumes of high-quality, pretreated data. Data pipelines help ensure models get the data they need to learn and make accurate predictions.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Ensuring data quality and governance: As data gets cleaned, validated, and standardized, data pipelines support greater confidence in data-driven decisions. They also enforce governance rules for compliance and security.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Improving operational efficiency: By integrating data from various systems, pipelines provide a holistic view of operations, automating workflows and flagging issues in real time.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Facilitating data democratization: Pipelines make data accessible and understandable to more people within an organization, empowering more teams to make informed decisions by connecting data sources to decision-makers.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Without strong data pipelines, organizations can fly blind—making decisions based on intuition rather than evidence.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h2><span class="TextRun SCXW175138420 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW175138420 BCX0" data-ccp-parastyle="heading 2">What are the core components of a data pipeline?</span></span></h2>
<p><span class="TextRun SCXW11624762 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW11624762 BCX0" data-ccp-parastyle="Normal (Web)">Every data pipeline is made up of five essential components. Understanding these elements reveals how data flows and transforms from its original source to its </span><span class="NormalTextRun AdvancedProofingIssueV2Themed SCXW11624762 BCX0" data-ccp-parastyle="Normal (Web)">final destination</span><span class="NormalTextRun SCXW11624762 BCX0" data-ccp-parastyle="Normal (Web)">.</span></span><span class="EOP Selected SCXW11624762 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3><span class="TextRun SCXW46434354 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW46434354 BCX0" data-ccp-parastyle="heading 3">1. Source: Where your data lives</span></span></h3>
<p><span data-contrast="auto">Source is the origin point of your data—the starting line of the pipeline. The type of source determines how the data will be extracted. Common data sources include:</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<ul>
<li><span data-contrast="auto">Databases: relational (e.g., MySQL) and NoSQL (e.g., MongoDB)</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Applications: CRM systems (e.g., Salesforce), ERPs (e.g., SAP), marketing automation platforms</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">APIs: third-party services, social media platforms, public data feeds</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Files: CSVs, JSON, XML, Parquet, Avro—often stored in cloud storage (e.g., Azure Blob)</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Streaming data: real-time event streams from IoT devices, website clicks, financial transactions</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Logs: system logs, web server logs, application logs</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
</ul>
<h3><span class="TextRun SCXW96349524 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW96349524 BCX0" data-ccp-parastyle="heading 3">2. Extraction: Getting your data out</span></span></h3>
<p><span class="TextRun SCXW192058207 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW192058207 BCX0" data-ccp-parastyle="Normal (Web)">Extraction is the step where data is pulled from its original source—often involving different file types, formats, and sometimes unstable or slow source connections. The goal of extraction is to get a raw copy of the data without altering the source system.</span></span></p>
<p><span class="TextRun SCXW40638522 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW40638522 BCX0" data-ccp-parastyle="Normal (Web)">Three common extraction methods:</span></span><span class="EOP Selected SCXW40638522 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<ul>
<li><span data-contrast="auto">Batch extraction: Data is pulled in chunks at scheduled intervals (e.g., nightly, hourly). Used for data that doesn&#8217;t change frequently or where immediate updates aren&#8217;t critical.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Incremental extraction: Only new or changed data since the last pipeline run is extracted. Faster than full extraction, but requires change-detection techniques like timestamps, version numbers, or Change Data Capture (CDC).</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Streaming extraction: Data is continuously pulled from sources as events occur, typically using message queues or event streaming platforms like Kafka or Kinesis.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
</ul>
<h3><span class="TextRun SCXW22719446 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW22719446 BCX0" data-ccp-parastyle="heading 3">3. Transformation: Cleaning and shaping your data</span></span><span class="EOP Selected SCXW22719446 BCX0" data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;335559738&quot;:160,&quot;335559739&quot;:80}"> </span></h3>
<p><span class="TextRun SCXW135518083 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW135518083 BCX0" data-ccp-parastyle="Normal (Web)">Transformation is usually the most complex part of a data pipeline. This is where the messiness of raw data gets cleaned and turned into actionable information. The goal is to ensure data quality, consistency, and suitability for the intended destination. Common transformation steps include:</span></span></p>
<ul>
<li><span data-contrast="auto">Cleaning: Removing duplicates, handling missing values, correcting errors</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Filtering: Selecting only relevant rows or columns</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Aggregating: Summarizing or categorizing data (e.g., total sales per day)</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Joining or merging: Combining data from multiple sources using common keys (e.g., joining customer data with order data)</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Standardizing or normalizing: Ensuring consistent data types, formats, and units (e.g., standardizing currency codes)</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Enriching: Adding new data points by looking up external information or deriving new features (e.g., adding geographical data based on an IP address)</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Structuring: Converting unstructured or semi-structured data into a structured format</span></li>
</ul>
<h3><span class="TextRun SCXW15967525 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW15967525 BCX0" data-ccp-parastyle="heading 3">4. Loading: Delivering your data</span></span></h3>
<p><span class="TextRun SCXW245129577 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW245129577 BCX0" data-ccp-parastyle="Normal (Web)">Once extracted and transformed, data is loaded into the system where it will be used—a database, data warehouse, or analytics platform—so applications, reports, and analytics tools can access it. Three common loading strategies:</span></span></p>
<ul>
<li><span data-contrast="auto">Full load: The entire destination table or dataset is overwritten with new, transformed data. Simpler to implement, but resource-intensive for large datasets.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Incremental load: Only new or changed records are appended to the destination. More efficient, but requires diligent management of data updates and deletions.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Streaming load: Data is continuously loaded as it arrives, often into specialized real-time databases or analytical engines.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
</ul>
<h3><span class="TextRun SCXW192782379 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW192782379 BCX0" data-ccp-parastyle="heading 3">5. Destination: Where your data rests</span></span></h3>
<p><span class="TextRun SCXW236999194 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW236999194 BCX0" data-ccp-parastyle="Normal (Web)">Destination is the final storage location where processed data is available for consumption by analysts, data scientists, and applications. Common destinations include:</span></span><span class="EOP Selected SCXW236999194 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<ul>
<li><span data-contrast="auto">Data warehouses: Optimized for complex queries and reporting on large volumes of historical data (e.g., Snowflake)</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Data lakes: Hold raw or semi-structured data at scale for advanced analytics and machine learning (e.g., Azure Data Lake Storage)</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Databases: Operational systems for everyday applications like websites or apps (e.g., MongoDB)</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Business intelligence (BI) tools: Software that turns data into dashboards and reports (e.g., Tableau)</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">File storage: Simple storage for archiving or later processing</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
</ul>
<p><span data-contrast="auto">Note: loading and destination are conceptually distinct but closely related in practice. Loading is the action of writing processed data into a system; destination is the place where that data ends up and stays until it&#8217;s needed. Loading is like putting groceries into the fridge. Destination is the actual fridge.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Bringing it all together—an e-commerce scenario: Source (CRM database) → Extract (SQL query for new customer orders) → Transform (clean addresses, calculate total order value, join with product details) → Load (insert into data warehouse) → Destination (data warehouse for reporting). The entire process runs automatically on a schedule, ensuring a continuous flow of refined information.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h2><span class="TextRun SCXW83207294 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW83207294 BCX0" data-ccp-parastyle="heading 2">What are the three types of data pipelines?</span></span></h2>
<p><span class="TextRun SCXW205833320 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW205833320 BCX0" data-ccp-parastyle="Normal (Web)">Just as there are </span><span class="NormalTextRun SCXW205833320 BCX0" data-ccp-parastyle="Normal (Web)">different ways</span><span class="NormalTextRun SCXW205833320 BCX0" data-ccp-parastyle="Normal (Web)"> to transport goods, there are </span><span class="NormalTextRun SCXW205833320 BCX0" data-ccp-parastyle="Normal (Web)">different types</span><span class="NormalTextRun SCXW205833320 BCX0" data-ccp-parastyle="Normal (Web)"> of data pipelines—each optimized for specific needs around speed, volume, and complexity.</span></span><span class="EOP Selected SCXW205833320 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3><span class="TextRun SCXW105155355 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW105155355 BCX0" data-ccp-parastyle="heading 3">Batch processing: the daily shuttle</span></span></h3>
<p><span class="TextRun SCXW27358944 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW27358944 BCX0" data-ccp-parastyle="Normal (Web)">Batch processing works like a commuter train on a defined schedule. It picks up a large group of passengers (data) at scheduled times and delivers them to their destination. Data is collected over a period, then processed as a single, large batch.</span></span></p>
<ul>
<li><span data-contrast="auto">Characteristics: High latency (data may be hours or days old); processes large volumes efficiently; often scheduled during off-peak hours</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Use cases: Nightly reports, monthly financial summaries, loading historical data into a data warehouse, running complex analytical jobs that don&#8217;t require immediate results</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
</ul>
<h3><span class="TextRun SCXW217176488 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW217176488 BCX0" data-ccp-parastyle="heading 3">Real-time streaming: the instant delivery service</span></span></h3>
<p><span class="TextRun SCXW17467150 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW17467150 BCX0" data-ccp-parastyle="Normal (Web)">Real-time streaming works like an instant delivery service. As soon as a package (data event) is created, it&#8217;s picked up, processed almost immediately, and delivered to its destination with minimal delay.</span></span></p>
<ul>
<li><span data-contrast="auto">Characteristics: Low latency (data is typically seconds or milliseconds old); handles continuous streams of individual events; requires infrastructure optimized for speed</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Use cases: Fraud detection, real-time personalized recommendations, IoT sensor data analysis, monitoring system health, live dashboards</span></li>
</ul>
<h3><span class="TextRun SCXW205110269 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW205110269 BCX0" data-ccp-parastyle="heading 3">Hybrid approaches: the best of both worlds</span></span></h3>
<p><span class="TextRun SCXW16224768 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW16224768 BCX0" data-ccp-parastyle="Normal (Web)">Many organizations combine batch and streaming pipelines. Two common hybrid patterns:</span></span></p>
<ul>
<li><span data-contrast="auto">Lambda architecture: Uses separate batch and streaming layers. The streaming layer provides real-time views; the batch layer processes historical data for accuracy and completeness. Results from both are then merged.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Kappa architecture: A simpler approach that handles both real-time and historical processing using a single streaming engine, often by replaying streams.</span></li>
</ul>
<p><span class="TextRun SCXW61867770 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW61867770 BCX0" data-ccp-parastyle="Normal (Web)">Choosing the right pipeline type depends entirely on your business requirements for data freshness, volume, and complexity.</span></span><span class="EOP Selected SCXW61867770 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h2><span class="TextRun SCXW183454366 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW183454366 BCX0" data-ccp-parastyle="heading 2">What are the main challenges in data pipeline management?</span></span></h2>
<p><span data-contrast="auto">Building a data pipeline is one thing; keeping it running smoothly and reliably is another. Here are the most common data pipeline management challenges—and the best practices for each.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Ensuring data quality: Data can be inconsistent, incomplete, or incorrect at the source, leading to garbage in, garbage out. Best practices: implement data validation rules at every stage; use data profiling tools to understand data characteristics; create data quality checks within transformation steps; use data observability platforms to detect anomalies early.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Scalability and performance: As data volumes grow or requirements shift to real-time, pipelines can become slow or break entirely. Best practices: design for scalability from the outset; use distributed processing frameworks (e.g., Apache Spark); use cloud-native services that scale automatically; implement incremental loading strategies; optimize queries and transformation logic.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Security and compliance: Data pipelines handle sensitive information, requiring stringent security and compliance measures. Best practices: encrypt data at rest and in transit; implement strong access controls (least privilege); audit data access and changes; redact sensitive data during transformation where necessary.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Monitoring and alerting: Without proper monitoring, pipeline failures or data quality issues can go undetected and impact downstream applications. Best practices: implement comprehensive monitoring for pipeline health, performance metrics, and data quality metrics; set up automated alerts for critical failures, latency breaches, or data anomalies; use dashboards for operational visibility.</span></p>
<h2><span class="TextRun SCXW21357821 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW21357821 BCX0" data-ccp-parastyle="heading 2">What are the top data pipeline use cases?</span></span></h2>
<p><span class="TextRun SCXW136485918 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW136485918 BCX0" data-ccp-parastyle="Normal (Web)">Data pipelines are versatile, powering applications across industries. Seven key examples of how pipelines transform raw data into actionable insights:</span></span></p>
<ul>
<li><span data-contrast="auto">Business intelligence and reporting: Aggregating sales data, customer demographics, and marketing spend into a data warehouse for daily, weekly, or monthly reports and dashboards that guide strategic decisions</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Customer 360-degree view: Combining data from CRM, sales, support, and marketing platforms to create a holistic profile of each customer, enabling personalized experiences and targeted campaigns</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Fraud detection: Ingesting real-time financial transactions, social media activity, and user behavior to identify suspicious patterns and instantly flag potential fraud</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">IoT analytics: Collecting streams of data from sensors (e.g., factory machines, smart city devices) to monitor performance, predict maintenance needs, and optimize operations</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Personalized recommendations: Processing user browsing history, purchase data, and demographic information to power content recommendations on streaming platforms</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Log analytics: Consolidating logs from applications and servers to monitor system health, troubleshoot issues, and detect security threats in real time</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">ML model training: Preparing, cleaning, and feeding large datasets to machine learning models for tasks like image recognition, natural language processing, or predictive analytics</span></li>
</ul>
<h2><span class="TextRun SCXW57361826 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW57361826 BCX0" data-ccp-parastyle="heading 2">What makes a data pipeline modern?</span></span></h2>
<p><span class="TextRun SCXW246538843 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW246538843 BCX0" data-ccp-parastyle="Normal (Web)">Modern data pipelines share several key characteristics that distinguish them from traditional approaches:</span></span></p>
<ul>
<li><span data-contrast="auto">Cloud-native and serverless: Modern pipelines use cloud services (e.g., AWS, Azure) that scale automatically and reduce operational overhead</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">ELT-first approach: Instead of transforming data before loading, modern pipelines often load raw data into a cloud data warehouse (e.g., Snowflake, BigQuery) and transform it within the warehouse using SQL—leveraging the destination&#8217;s compute power and enabling greater flexibility</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Data lake integration: Modern pipelines frequently integrate with data lakes to store vast amounts of raw, multi-structured data for future use, advanced analytics, and machine learning</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Real-time capabilities: Streaming technologies like Apache Kafka and Amazon Kinesis increasingly power modern pipelines for immediate insights</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Orchestration and automation: Tools like Apache Airflow or cloud-native orchestrators automate scheduling, manage dependencies, and monitor pipeline health</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Data observability: Modern pipelines go beyond basic monitoring to actively track the health, quality, and lineage of data—detecting anomalies and ensuring data trustworthiness</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Data governance and security by design: Security, privacy, and compliance are built into the pipeline architecture from the start, not added as an afterthought</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Flexibility and agility: Modular components make modern pipelines easier to adapt to new data sources and changing business requirements</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
</ul>
<h2><span class="TextRun SCXW234200500 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW234200500 BCX0" data-ccp-parastyle="heading 2">How do traditional and modern data pipelines compare?</span></span><span class="EOP Selected SCXW234200500 BCX0" data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;335559738&quot;:160,&quot;335559739&quot;:80}"> </span></h2>
<p><span class="TextRun SCXW105980290 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW105980290 BCX0" data-ccp-parastyle="Normal (Web)">Modern data pipelines are designed to be more agile, scalable, cost-effective, and resilient than traditional approaches. </span><span class="NormalTextRun SCXW105980290 BCX0" data-ccp-parastyle="Normal (Web)">Here&#8217;s</span><span class="NormalTextRun SCXW105980290 BCX0" data-ccp-parastyle="Normal (Web)"> a side-by-side comparison:</span></span></p>
<table data-tablestyle="MsoTable15Plain4" data-tablelook="1184" aria-rowcount="8">
<tbody>
<tr aria-rowindex="1">
<td data-celllook="0"><b><span data-contrast="auto">Feature </span></b><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><b><span data-contrast="auto">Traditional Data Pipeline</span></b><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><b><span data-contrast="auto">Modern Data Pipeline</span></b><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
</tr>
<tr aria-rowindex="2">
<td data-celllook="0"><b><span data-contrast="auto">Scalability</span></b><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><span data-contrast="auto">Limited by fixed resources and batch processing constraints</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><span class="TextRun SCXW132747783 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW132747783 BCX0">Highly scalable and elastic, using cloud infrastructure to adjust resources automatically</span></span></td>
</tr>
<tr aria-rowindex="3">
<td data-celllook="0"><b><span data-contrast="auto">Processing</span></b><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><span data-contrast="auto">Primarily batch processing (e.g., hourly, daily)</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><span data-contrast="auto">Supports both batch and continuous, real-time processing</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
</tr>
<tr aria-rowindex="4">
<td data-celllook="0"><b><span data-contrast="auto">Flexibility</span></b><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><span data-contrast="auto">Less flexible; requires significant manual adjustments for changes</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><span data-contrast="auto">More flexible and adaptable; uses metadata to handle changes automatically</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
</tr>
<tr aria-rowindex="5">
<td data-celllook="0"><b><span data-contrast="auto">Infrastructure</span></b><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><span class="TextRun SCXW141839912 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW141839912 BCX0">Traditional, monolithic, on-premises systems</span></span></td>
<td data-celllook="0"><span class="TextRun SCXW242721589 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW242721589 BCX0">Cloud-native and microservices-based, with independent compute resources</span></span></td>
</tr>
<tr aria-rowindex="6">
<td data-celllook="0"><b><span data-contrast="auto">Automation</span></b><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><span data-contrast="auto">Limited automation</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><span class="TextRun SCXW150699946 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW150699946 BCX0">High automation, including automated restarts and retries</span></span></td>
</tr>
<tr aria-rowindex="7">
<td data-celllook="0"><b><span data-contrast="auto">Data access</span></b><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><span data-contrast="auto">Data access can be restricted</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><span data-contrast="auto">Democratizes data access and enables self-service management</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
</tr>
<tr aria-rowindex="8">
<td data-celllook="0"><b><span data-contrast="auto">Real-time capabilities</span></b><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><span data-contrast="auto">Lower latency due to batching; not typically real-time</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
<td data-celllook="0"><span data-contrast="auto">Low latency with options for real-time processing and immediate data availability</span><span data-ccp-props="{&quot;201341983&quot;:2,&quot;335559740&quot;:330}"> </span></td>
</tr>
</tbody>
</table>
<h2><span class="TextRun SCXW114076102 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW114076102 BCX0" data-ccp-parastyle="heading 2">What are the most common data pipeline tools?</span></span></h2>
<p><span data-contrast="auto">The data pipeline tool landscape is broad and evolving. Four categories to consider:</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Data integration platforms: Comprehensive ETL/ELT tools, often with visual interfaces and pre-built connectors. Examples: Talend, Informatica. A good fit for teams that want an end-to-end solution without heavy coding; businesses with multiple data sources needing pre-built connectors; and organizations prioritizing ease of use and quick deployment.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Cloud-native services: Major cloud providers offer services specifically designed for scalable data pipelines. Examples: Amazon Kinesis, Google BigQuery. A good fit for companies already invested in a specific cloud ecosystem; teams needing scalable, cost-effective solutions for batch and streaming; and use cases requiring tight integration with other cloud services.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Open-source frameworks: Flexible, developer-friendly options for orchestration and processing. Examples: Apache Airflow, Apache Kafka. A good fit for engineering teams with strong technical skills; organizations wanting maximum flexibility and control; and scenarios with custom requirements or large-scale data processing needs.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Enterprise workflow orchestration platforms like Control-M are ideal for teams that need to <a href="/it-solutions/control-m-integrations/gcp-big-query.html">orchestrate BigQuery data pipelines</a> alongside mainframe, cloud, and on-premises workloads — providing SLA management and audit capabilities that native GCP scheduling does not offer. A good fit for large enterprises with complex, mission-critical workflows; businesses needing robust scheduling, compliance, and audit capabilities; and teams managing cross-platform jobs spanning mainframe, cloud, and on-premises systems with high reliability requirements.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">The right tool depends on your budget, team skill set, data volumes, and real-time requirements.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h2><span class="TextRun SCXW212447222 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW212447222 BCX0" data-ccp-parastyle="heading 2">5 key data pipeline takeaways</span></span></h2>
<p><span class="TextRun SCXW157260854 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW157260854 BCX0" data-ccp-parastyle="Normal (Web)">Whether </span><span class="NormalTextRun SCXW157260854 BCX0" data-ccp-parastyle="Normal (Web)">you&#8217;re</span><span class="NormalTextRun SCXW157260854 BCX0" data-ccp-parastyle="Normal (Web)"> learning about data pipelines for the first time or refreshing on the fundamentals, here are the most important points to carry forward:</span></span><span class="EOP Selected SCXW157260854 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3><span class="TextRun SCXW56389995 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW56389995 BCX0" data-ccp-parastyle="heading 3">1. Understand the pipeline lifecycle</span></span></h3>
<p><span class="TextRun SCXW71223408 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW71223408 BCX0" data-ccp-parastyle="Normal (Web)">A data pipeline </span><span class="NormalTextRun SCXW71223408 BCX0" data-ccp-parastyle="Normal (Web)">isn&#8217;t</span><span class="NormalTextRun SCXW71223408 BCX0" data-ccp-parastyle="Normal (Web)"> </span><span class="NormalTextRun SCXW71223408 BCX0" data-ccp-parastyle="Normal (Web)">just about moving</span><span class="NormalTextRun SCXW71223408 BCX0" data-ccp-parastyle="Normal (Web)"> data—it involves extraction, transformation, loading, orchestration, monitoring, and governance.</span></span></p>
<h3><span class="TextRun SCXW25048615 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW25048615 BCX0" data-ccp-parastyle="heading 3">2. Orchestration is key</span></span></h3>
<p><span class="TextRun SCXW65510838 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW65510838 BCX0" data-ccp-parastyle="Normal (Web)">Orchestration ensures repeatability, scalability, and observability across the entire data pipeline.</span></span><span class="EOP Selected SCXW65510838 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3><span class="TextRun SCXW45674670 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW45674670 BCX0" data-ccp-parastyle="heading 3">3. Embrace automation and CI/CD</span></span></h3>
<p><span class="TextRun SCXW156576247 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW156576247 BCX0" data-ccp-parastyle="Normal (Web)">Integrating data pipelines into CI/CD workflows enables faster, safer changes. </span></span><a class="Hyperlink SCXW156576247 BCX0" href="/info/dataops.html" rel="noreferrer noopener"><span class="TextRun Underlined SCXW156576247 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW156576247 BCX0" data-ccp-charstyle="Hyperlink">DataOps</span></span></a><span class="TextRun SCXW156576247 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW156576247 BCX0" data-ccp-parastyle="Normal (Web)"> applies </span></span><a class="Hyperlink SCXW156576247 BCX0" href="/info/devops.html" rel="noreferrer noopener"><span class="TextRun Underlined SCXW156576247 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW156576247 BCX0" data-ccp-charstyle="Hyperlink">DevOps</span></span></a><span class="TextRun SCXW156576247 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW156576247 BCX0" data-ccp-parastyle="Normal (Web)"> principles to data: automated testing, deployment, and version control for pipelines.</span></span></p>
<h3>4. Prioritize data quality and monitoring</h3>
<p><span data-contrast="auto">Pipelines can fail silently if data quality isn&#8217;t checked. Implement validation, anomaly detection, and alerts to catch issues early. Observability is critical for trust and compliance.</span></p>
<h3>5. Design for scalability and flexibility</h3>
<p><span data-contrast="auto">Modern data pipelines must handle batch and streaming, adapt to schema changes, and scale with data growth. Cloud-native and modular architectures are essential for agility.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Bonus tip: Learn the full ecosystem—ETL tools, orchestration frameworks, cloud services—and how they fit together to build pipelines that serve your organization at scale.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h2><span class="TextRun SCXW64168813 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW64168813 BCX0" data-ccp-parastyle="heading 2">Frequently asked questions about data pipelines</span></span></h2>
<p><strong>How is a data pipeline different from a data warehouse?</strong><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">A data pipeline is the process that moves and transforms data; a data warehouse is a destination where processed data is stored for analysis. Data pipelines feed data warehouses—the pipeline handles transport and preparation, while the warehouse provides the storage and query environment. The two are complementary, not interchangeable.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What skills do you need to build a data pipeline? </strong></p>
<p><span data-contrast="auto">Building a data pipeline typically requires proficiency in SQL and at least one programming language (Python and Scala are common), familiarity with data transformation concepts (ETL/ELT), and working knowledge of at least one orchestration or workflow tool. Cloud platform experience (AWS, Azure, or GCP) is increasingly essential for modern data pipeline work.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>How do you test a data pipeline? </strong></p>
<p><span data-contrast="auto">Data pipeline testing involves validating data at multiple stages: confirming that extraction produces complete, unaltered source data; verifying that transformation logic produces expected outputs; and checking that loaded data matches expected row counts, data types, and values at the destination. Automated testing frameworks integrated into CI/CD pipelines can run these checks continuously and catch regressions early.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What is data pipeline latency? </strong></p>
<p><span data-contrast="auto">Data pipeline latency is the time between a data event occurring at the source and that data becoming available at the destination. Batch pipelines have high latency (hours or days); real-time streaming pipelines have low latency (seconds or milliseconds). Acceptable latency depends on the use case—fraud detection requires near-zero latency, while monthly financial reporting can tolerate daily batch processing.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>How does BMC Control-M help manage data pipelines?</strong><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">BMC Control-M is an enterprise workflow orchestration platform that automates, schedules, and monitors complex data pipelines across environments—including mainframe, cloud, and on-premises systems. Control-M provides centralized visibility, dependency management, and audit capabilities, making it well-suited for mission-critical pipelines that require high reliability, compliance, and cross-platform coordination.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><em>The views and opinions expressed in this post are those of the author and do not necessarily reflect the official position of BMC. </em></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>How to orchestrate a data pipeline on Google Cloud with Control-M from BMC</title>
		<link>https://blogs.bmc.com/orchestrate-a-data-pipeline/</link>
		
		<dc:creator><![CDATA[Joe Goldberg]]></dc:creator>
		<pubDate>Mon, 30 Mar 2026 16:17:22 +0000</pubDate>
				<category><![CDATA[Machine Learning & Big Data Blog]]></category>
		<guid isPermaLink="false">https://blogs.bmc.com/?p=52266</guid>

					<description><![CDATA[<img width="810" height="405" src="https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-1024x512.jpg.optimal.jpg" class="attachment-large size-large wp-post-image" alt="SADA" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-1024x512.jpg.optimal.jpg 1024w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-300x150.jpg.optimal.jpg 300w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-768x384.jpg.optimal.jpg 768w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-810x405.jpg.optimal.jpg 810w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-1140x570.jpg.optimal.jpg 1140w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-24x12.jpg.optimal.jpg 24w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-36x18.jpg.optimal.jpg 36w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-48x24.jpg.optimal.jpg 48w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline.jpg.optimal.jpg 1400w" sizes="auto, (max-width: 810px) 100vw, 810px" />Control-M from BMC enables teams to orchestrate data pipelines on Google Cloud by defining, scheduling, and monitoring workflows across services like Cloud Storage, Dataflow, and BigQuery—all from a single interface. This article walks through the five key orchestration challenges, the Google Cloud services involved, and a real-world credit-card fraud detection example that puts Control-M into action.  […]]]></description>
										<content:encoded><![CDATA[<img width="810" height="405" src="https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-1024x512.jpg.optimal.jpg" class="attachment-large size-large wp-post-image" alt="SADA" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-1024x512.jpg.optimal.jpg 1024w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-300x150.jpg.optimal.jpg 300w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-768x384.jpg.optimal.jpg 768w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-810x405.jpg.optimal.jpg 810w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-1140x570.jpg.optimal.jpg 1140w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-24x12.jpg.optimal.jpg 24w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-36x18.jpg.optimal.jpg 36w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline-48x24.jpg.optimal.jpg 48w, https://s7280.pcdn.co/wp-content/uploads/2022/09/orchestrate-a-data-pipeline.jpg.optimal.jpg 1400w" sizes="auto, (max-width: 810px) 100vw, 810px" /><p><span class="TextRun SCXW208330064 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW208330064 BCX0" data-ccp-parastyle="Normal (Web)">Control-M from BMC enables teams to orchestrate data pipelines on Google Cloud by defining, scheduling, and monitoring workflows across services like Cloud Storage, Dataflow, and </span><span class="NormalTextRun SpellingErrorV2Themed SCXW208330064 BCX0" data-ccp-parastyle="Normal (Web)">BigQuery</span><span class="NormalTextRun SCXW208330064 BCX0" data-ccp-parastyle="Normal (Web)">—all from a single interface. This article walks through the five key orchestration challenges, the Google Cloud services involved, and a real-world credit-card fraud detection example that puts Control-M into action.</span></span><span class="EOP Selected SCXW208330064 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span class="TextRun SCXW35380159 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW35380159 BCX0" data-ccp-parastyle="Normal (Web)">The </span></span><a class="Hyperlink SCXW35380159 BCX0" href="https://cloud.google.com/" target="_blank" rel="noreferrer noopener"><span class="TextRun Underlined SCXW35380159 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW35380159 BCX0" data-ccp-charstyle="Hyperlink">Google Cloud Platform</span></span></a><span class="TextRun SCXW35380159 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW35380159 BCX0" data-ccp-parastyle="Normal (Web)"> is designed specifically to accommodate organizations in a variety of positions along their cloud services journey, from large-scale machine learning (ML) and data analysis to services tailored to SMBs to hybrid-cloud solutions for customers that want to use services from more than one cloud provider. When </span></span><a class="Hyperlink SCXW35380159 BCX0" href="https://www.bmc.com/" target="_blank" rel="noreferrer noopener"><span class="TextRun Underlined SCXW35380159 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW35380159 BCX0" data-ccp-charstyle="Hyperlink">BMC</span></span></a><span class="TextRun SCXW35380159 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW35380159 BCX0" data-ccp-parastyle="Normal (Web)"> was migrating our Control-M application to this cloud ecosystem, we had to be very thoughtful about how we managed this </span><span class="NormalTextRun SCXW35380159 BCX0" data-ccp-parastyle="Normal (Web)">change. The </span></span><a class="Hyperlink SCXW35380159 BCX0" href="https://sada.com/" target="_blank" rel="noreferrer noopener"><span class="TextRun Underlined SCXW35380159 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW35380159 BCX0" data-ccp-charstyle="Hyperlink">SADA</span></span></a><span class="TextRun SCXW35380159 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW35380159 BCX0" data-ccp-parastyle="Normal (Web)"> engineering team worked alongside the BMC team to ensure that we had a seamless integration for our customers.</span></span><span class="EOP SCXW35380159 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span class="TextRun SCXW254758208 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW254758208 BCX0" data-ccp-parastyle="Normal (Web)">SADA supported this project by providing an inventory of the Google Cloud configuration options, decisions, and recommendations to enable the data platform foundation deployment, collaborated with BMC on the implementation planning, provided automation templates, and designed the Google Cloud architecture for the relevant managed services on the Google Cloud Platform.</span></span></p>
<p><span class="TextRun SCXW242857786 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW242857786 BCX0" data-ccp-parastyle="Normal (Web)">In this article, we will discuss </span><span class="NormalTextRun AdvancedProofingIssueV2Themed SCXW242857786 BCX0" data-ccp-parastyle="Normal (Web)">the end result</span><span class="NormalTextRun SCXW242857786 BCX0" data-ccp-parastyle="Normal (Web)"> of this </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW242857786 BCX0" data-ccp-parastyle="Normal (Web)">work, and</span><span class="NormalTextRun SCXW242857786 BCX0" data-ccp-parastyle="Normal (Web)"> look at an example using a credit-card fraud detection process to show how you can use Control-M to orchestrate a data pipeline seamlessly in Google Cloud.</span></span></p>
<h2><span class="TextRun SCXW37534474 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW37534474 BCX0" data-ccp-parastyle="heading 2">What are the five challenges of orchestrating an ML data pipeline?</span></span></h2>
<p><span class="TextRun SCXW87668284 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW87668284 BCX0" data-ccp-parastyle="Normal (Web)">Orchestrating a machine learning data pipeline on Google Cloud involves five primary challenges that teams must address before workflows can run reliably at scale.</span></span></p>
<h3>Understand the workflow</h3>
<p><span class="TextRun SCXW144105476 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW144105476 BCX0" data-ccp-parastyle="Normal (Web)">Examine all dependencies and any </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW144105476 BCX0" data-ccp-parastyle="Normal (Web)">decision</span><span class="NormalTextRun SCXW144105476 BCX0" data-ccp-parastyle="Normal (Web)"> trees. </span><span class="NormalTextRun SCXW144105476 BCX0" data-ccp-parastyle="Normal (Web)">For example, if data ingestion is successful, proceed down this path; if it is not successful, proceed down that path.</span></span></p>
<h3>Understand the teams</h3>
<p>If multiple teams are involved in the workflow, each needs to have a way to define their workflow using a standard interface, and to be able to merge their workflows to make up the pipeline.</p>
<h3>Follow standards</h3>
<p>Teams should use repeatable standards and conventions when building workflows. This avoids having multiple jobs with identical names. Each step should also have a meaningful description to help clarify its purpose in the event of a failure.</p>
<h3>Minimize the number of tools required</h3>
<p>Use a single tool for visualization and interaction with the pipeline (and dependencies). Visualization is important during the definition stage since it’s hard to manage something that you can’t see. This is even more important when the pipeline is running.</p>
<h3>Include built-in error handling capabilities</h3>
<p><span class="TextRun SCXW143694447 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW143694447 BCX0" data-ccp-parastyle="Normal (Web)">It&#8217;s</span><span class="NormalTextRun SCXW143694447 BCX0" data-ccp-parastyle="Normal (Web)"> important to understand how errors can </span><span class="NormalTextRun SCXW143694447 BCX0" data-ccp-parastyle="Normal (Web)">impact</span><span class="NormalTextRun SCXW143694447 BCX0" data-ccp-parastyle="Normal (Web)"> downstream jobs in the workflow or the business service level agreement (SLA). Failure of a job should not halt the pipeline altogether and require human interaction. Criteria can be used to </span><span class="NormalTextRun SCXW143694447 BCX0" data-ccp-parastyle="Normal (Web)">determine</span><span class="NormalTextRun SCXW143694447 BCX0" data-ccp-parastyle="Normal (Web)"> if a failed job can be restarted automatically or whether a human must be contacted to evaluate the failure—if, for instance, there are a certain number of failures involving the same error.</span></span></p>
<h2><span class="TextRun SCXW155267346 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW155267346 BCX0" data-ccp-parastyle="heading 2">How did BMC and SADA meet these orchestration challenges?</span></span></h2>
<p><span class="TextRun SCXW266037518 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW266037518 BCX0" data-ccp-parastyle="Normal (Web)">Addressing these challenges required a solid foundation and presented opportunities for collaboration. BMC and SADA aligned using the SADA POWER line of services to </span><span class="NormalTextRun SCXW266037518 BCX0" data-ccp-parastyle="Normal (Web)">establish</span><span class="NormalTextRun SCXW266037518 BCX0" data-ccp-parastyle="Normal (Web)"> the data platform foundation. Some notable elements in this technical alignment included work by SADA to:</span></span></p>
<ul>
<li><span data-contrast="auto">Apply industry expertise to expedite BMC&#8217;s development efforts.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Establish a best practices baseline around data pipelines and the tools to orchestrate them.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Conduct collaborative sessions to understand BMC&#8217;s technical needs and provide solutions that the BMC team could integrate and then expand upon.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
</ul>
<p><span class="TextRun SCXW118623991 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW118623991 BCX0" data-ccp-parastyle="Normal (Web)">SADA&#8217;s Data Platform Foundation provided opportunities to </span><span class="NormalTextRun SCXW118623991 BCX0" data-ccp-parastyle="Normal (Web)">leverage</span><span class="NormalTextRun SCXW118623991 BCX0" data-ccp-parastyle="Normal (Web)"> Google Cloud services to </span><span class="NormalTextRun SCXW118623991 BCX0" data-ccp-parastyle="Normal (Web)">accomplish</span><span class="NormalTextRun SCXW118623991 BCX0" data-ccp-parastyle="Normal (Web)"> the complex analytics required of an application like Control-M. The BMC and SADA teams worked together to </span><span class="NormalTextRun SCXW118623991 BCX0" data-ccp-parastyle="Normal (Web)">establish</span><span class="NormalTextRun SCXW118623991 BCX0" data-ccp-parastyle="Normal (Web)"> </span><span class="NormalTextRun SCXW118623991 BCX0" data-ccp-parastyle="Normal (Web)">a strong foundation</span><span class="NormalTextRun SCXW118623991 BCX0" data-ccp-parastyle="Normal (Web)"> through:</span></span></p>
<ul>
<li><span data-contrast="auto">Selecting data and storage locations in Google Cloud Storage.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Utilizing the advantages provided by Pub/Sub to streamline the analytics and data integration pipelines.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Having thorough discussions around the extract, transform, and load (ETL) processes to truly understand the end state of the data.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Using BigQuery and writing analytic queries.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Understanding the importance of automation, replicability of processes, and monitoring performance in establishing a system that is scalable and flexible.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Using Data Studio to create a visualization dashboard to provide the necessary business insights.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
</ul>
<h2><span class="TextRun SCXW99446985 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW99446985 BCX0" data-ccp-parastyle="heading 2">How does fraud detection illustrate data pipeline orchestration on Google Cloud?</span></span></h2>
<p><span class="TextRun SCXW164576330 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW164576330 BCX0" data-ccp-parastyle="Normal (Web)">Credit-card fraud detection is a practical, real-world example of how Control-M can orchestrate a complex ML data pipeline on Google Cloud—combining real-time and batch processes across multiple services.</span></span></p>
<p><span class="TextRun SCXW167563580 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW167563580 BCX0" data-ccp-parastyle="Normal (Web)">Digital transactions have been increasing steadily for years, and the accelerating adoption of digital payments by businesses and consumers has brought with it increased fraud and operational risks. With fraudsters improving their techniques, companies are relying on ML to build resilient and efficient fraud detection systems. Since fraud constantly evolves, detection systems must be able to </span><span class="NormalTextRun SCXW167563580 BCX0" data-ccp-parastyle="Normal (Web)">identify</span><span class="NormalTextRun SCXW167563580 BCX0" data-ccp-parastyle="Normal (Web)"> new types of fraud by detecting anomalies that are seen for the first time—making fraud detection a perpetual task that requires constant diligence and innovation.</span></span></p>
<p><span class="TextRun SCXW177272528 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW177272528 BCX0" data-ccp-parastyle="Normal (Web)">Common types of financial fraud that customers work to prevent with this application include:</span></span></p>
<ul>
<li>Stolen/fake credit card fraud: Transactions made using fake cards, or cards belonging to someone else.</li>
<li>ATM fraud: Cash withdrawals using someone else’s card.</li>
</ul>
<p>Fraud detection is composed of both real-time and batch processes. The real-time process is responsible for denying a transaction and possibly placing a hold on an account or credit card, thus preventing the fraud from occurring. It must respond quickly, sometimes at the cost of reduced accuracy.</p>
<p>To minimize false positives, which may upset or inconvenience customers, a batch phase is used to continuously fine-tune the detection model. After transactions are confirmed as valid or fraudulent, all recent events are input to the batch process on a regular cadence. This batch process then updates the training and scoring of the real-time model to keep real-time detection operating at peak accuracy. This batch process is the focus of this article.</p>
<h2><span class="TextRun SCXW28136590 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW28136590 BCX0" data-ccp-parastyle="heading 2">How can you try the demo system?</span></span></h2>
<p><span class="NormalTextRun SCXW140759508 BCX0" data-ccp-parastyle="Normal (Web)">SADA and BMC created a demonstration version of this </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW140759508 BCX0" data-ccp-parastyle="Normal (Web)">solution</span><span class="NormalTextRun SCXW140759508 BCX0" data-ccp-parastyle="Normal (Web)"> so you can experiment with it on Google Cloud. You can find all of the code, plus sample data, in</span> <a href="https://github.com/controlm/automation-api-community-solutions/tree/master/1-general-examples/use-case-gcp-fraud-detection">GitHub</a>.</p>
<p><span class="TextRun SCXW121347673 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW121347673 BCX0" data-ccp-parastyle="Normal (Web)">Resources included are:</span></span></p>
<ul>
<li>Kaggle datasets of transaction data, fraud status, and demographics</li>
<li>Queries</li>
<li>Schema</li>
<li>User-defined functions (UDFs)</li>
</ul>
<h2><span class="TextRun SCXW94439837 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW94439837 BCX0" data-ccp-parastyle="heading 2">How does the pipeline work?</span></span></h2>
<p>For each region in which the organization operates, transaction data is collected daily. Details collected include (but are not limited to):</p>
<ul>
<li>Transaction details: Describes each transaction, including the amount, item code, location, method of payment, and so on.</li>
<li>Personal details: Describes the name, address, age, and other details about the purchaser.</li>
</ul>
<p>This information is pulled from corporate data based on credit card information and real-time fraud detection that identifies which transactions were flagged as fraudulent.</p>
<p>New data arrives either as batch feeds or is dropped into Cloud Storage by Pub/Sub. This new data is then loaded into BigQuery by Dataflow jobs. Normalization and some data enrichment is performed by UDFs during the load process.</p>
<p>Once all the data preparation is complete, analytics are run against the combined new and historical data to test and rank fraud detection performance. The results are displayed in Data Studio dashboards.</p>
<div id="attachment_52267" style="width: 634px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" aria-describedby="caption-attachment-52267" class="wp-image-52267 size-full" src="https://s7280.pcdn.co/wp-content/uploads/2022/09/control-m-orchestration.png" alt="Control-M orchestration" width="624" height="305" srcset="https://s7280.pcdn.co/wp-content/uploads/2022/09/control-m-orchestration.png 624w, https://s7280.pcdn.co/wp-content/uploads/2022/09/control-m-orchestration-300x147.png 300w, https://s7280.pcdn.co/wp-content/uploads/2022/09/control-m-orchestration-24x12.png 24w, https://s7280.pcdn.co/wp-content/uploads/2022/09/control-m-orchestration-36x18.png 36w, https://s7280.pcdn.co/wp-content/uploads/2022/09/control-m-orchestration-48x23.png 48w" sizes="auto, (max-width: 624px) 100vw, 624px" /><p id="caption-attachment-52267" class="wp-caption-text">Figure 1: Control-M orchestration</p></div>
<h2><span class="TextRun SCXW30904278 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW30904278 BCX0" data-ccp-parastyle="heading 2">Which Google Cloud services power the pipeline?</span></span></h2>
<p><span class="TextRun SCXW221483902 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW221483902 BCX0" data-ccp-parastyle="Normal (Web)">Control-M orchestrates a coordinated set of Google Cloud services—Cloud Storage, Dataflow, </span><span class="NormalTextRun SpellingErrorV2Themed SCXW221483902 BCX0" data-ccp-parastyle="Normal (Web)">BigQuery</span><span class="NormalTextRun SCXW221483902 BCX0" data-ccp-parastyle="Normal (Web)">, and Data Studio—each handling a distinct stage of the data pipeline.</span></span></p>
<p><span class="TextRun SCXW243025447 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW243025447 BCX0" data-ccp-parastyle="Normal (Web)">Cloud Storage provides a common landing zone for all incoming data and </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW243025447 BCX0" data-ccp-parastyle="Normal (Web)">a consistent</span><span class="NormalTextRun SCXW243025447 BCX0" data-ccp-parastyle="Normal (Web)"> input for downstream processing. Dataflow is Google Cloud&#8217;s primary data integration tool.</span></span></p>
<p><span class="TextRun SCXW11574250 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW11574250 BCX0" data-ccp-parastyle="Normal (Web)">SADA and BMC selected </span><span class="NormalTextRun SpellingErrorV2Themed SCXW11574250 BCX0" data-ccp-parastyle="Normal (Web)">BigQuery</span><span class="NormalTextRun SCXW11574250 BCX0" data-ccp-parastyle="Normal (Web)"> for data processing. Earlier versions of this application used Hadoop, but while working with the team at SADA, we converted to </span><span class="NormalTextRun SpellingErrorV2Themed SCXW11574250 BCX0" data-ccp-parastyle="Normal (Web)">BigQuery</span><span class="NormalTextRun SCXW11574250 BCX0" data-ccp-parastyle="Normal (Web)"> as this is the recommended strategy from Google for sophisticated data warehouse or data lake applications. This choice also simplified setup by providing out-of-the-box integration with Cloud Dataflow. UDFs </span><span class="NormalTextRun SCXW11574250 BCX0" data-ccp-parastyle="Normal (Web)">provide</span><span class="NormalTextRun SCXW11574250 BCX0" data-ccp-parastyle="Normal (Web)"> a simple mechanism for manipulating data during the load process.</span></span></p>
<h2><span class="TextRun SCXW164216629 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW164216629 BCX0" data-ccp-parastyle="heading 2">What are the two ways to define pipeline workflows in Control-M?</span></span></h2>
<p><span class="TextRun SCXW30383976 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW30383976 BCX0" data-ccp-parastyle="Normal (Web)">Control-M supports two approaches for defining pipeline workflows, giving teams flexibility to work visually or programmatically.</span></span></p>
<h3>Using a graphical editor</h3>
<p>This provides the option of dragging and dropping the workflow steps into a workspace and connecting them.</p>
<h3>Use RESTful APIs</h3>
<p><span class="TextRun SCXW211862022 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW211862022 BCX0" data-ccp-parastyle="Normal (Web)">Define the workflows using a jobs-as-code method, then use JSON to integrate into a continuous integration/continuous delivery (CI/CD) toolchain. This method improves workflow management by flowing jobs through a pipeline of automated building, testing, and release. </span><span class="NormalTextRun SCXW211862022 BCX0" data-ccp-parastyle="Normal (Web)">Google Cloud provides a number of developer tools for CI/CD, including Cloud Build and Cloud Deploy.</span></span></p>
<h2><span class="TextRun SCXW196068014 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW196068014 BCX0" data-ccp-parastyle="heading 2">How are jobs defined in Control-M?</span></span></h2>
<p><span class="TextRun SCXW215157492 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW215157492 BCX0" data-ccp-parastyle="Normal (Web)">The basic Control-M execution unit is referred to as a job. There are </span><span class="NormalTextRun AdvancedProofingIssueV2Themed SCXW215157492 BCX0" data-ccp-parastyle="Normal (Web)">a number of</span><span class="NormalTextRun SCXW215157492 BCX0" data-ccp-parastyle="Normal (Web)"> attributes for every job, defined in JSON:</span></span></p>
<h3>Job type</h3>
<p>Options include script, command, file transfer, Dataflow, or BigQuery.</p>
<h3>Run location</h3>
<p>For instance, which host is running the job?</p>
<h3>Identity</h3>
<p>For example, is the job being “run as…” or run using a connection profile?</p>
<h3>Schedule</h3>
<p>Determines when to run the job and identifies relevant scheduling criteria.</p>
<h3>Dependencies</h3>
<p>This could be things like whether the job must finish by a certain time or output must arrive by a certain time or date.</p>
<p>Jobs are stored in folders and the attributes discussed above, along with any other instructions, are applied to all jobs in that folder.</p>
<p><span class="TextRun SCXW128757721 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW128757721 BCX0" data-ccp-parastyle="Normal (Web)">Below is an example of the JSON code that describes the workflow used in the fraud detection model ranking application. You can find the full JSON code in the </span></span><a class="Hyperlink SCXW128757721 BCX0" href="https://github.com/controlm/automation-api-community-solutions/tree/master/1-general-examples/use-case-gcp-fraud-detection" target="_blank" rel="noreferrer noopener"><span class="TextRun Underlined SCXW128757721 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW128757721 BCX0" data-ccp-charstyle="Hyperlink">Control-M Automation API Community Solutions GitHub repo</span></span></a><span class="TextRun SCXW128757721 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW128757721 BCX0" data-ccp-parastyle="Normal (Web)">. While there, you can also find solutions, the Control-M Automation API guide, and other code samples.</span></span></p>
<pre>{
"Defaults" : {
},
"jog-mc-gcp-fraud-detection": {"Type": "Folder",
"Comment" : "Update fraud history, run, train and score models",
"jog-gcs-download" : {"Type" : "Job:FileTransfer",…},
"jog-dflow-gcs-to-bq-fraud": {"Type": "Job:Google DataFlow",…},
"jog-dflow-gcs-to-bq-transactions": {"Type": “Job:Google DataFlow",…},
"jog-dflow-gcs-to-bq-personal": {"Type": "Job:Google DataFlow",…},
"jog-mc-bq-query": {"Type": "Job:Database:EmbeddedQuery", …},
"jog-mc-fm-service": {"Type": "Job:SLAManagement",…},
},
"flow00": {"Type":"Flow", "Sequence":[
"jog-gcs-download",
"jog-dflow-gcs-to-bq-fraud",
"jog-mc-bq-query",
"jog-mc-fm-service"]},
"flow01": {"Type":"Flow", "Sequence":[
"jog-gcs-download",
"jog-dflow-gcs-to-bq-transactions",
"jog-mc-bq-query", "jog-mc-fm-service"]},
"flow02": {"Type":"Flow", "Sequence":[
"jog-gcs-download",
"jog-dflow-gcs-to-bq-personal",
"jog-mc-bq-query",
"jog-mc-fm-service"]}

}
}
</pre>
<p>The jobs shown in this workflow correspond directly with the steps illustrated previously in Figure 1.</p>
<p>The workflow contains three fundamental sections:</p>
<p>Defaults. These are the functions that apply to the workflow. This could include details such as who to contact for job failures or standards for job naming or structure.</p>
<pre>{  "Defaults" : {"RunAs" : "ctmagent", "OrderMethod": "Manual", "Application" : 
       "multicloud", "SubApplication" : "jog-mc-fraud-modeling", 
      "Job" : {"SemQR": { "Type": "Resource:Semaphore", Quantity": "1"},
      "actionIfError" : {"Type": "If", "CompletionStatus":"NOTOK", "mailTeam": 
          {"Type": "Mail", "Message": "Job %%JOBNAME failed", "Subject": 
                 "Error occurred", "To": deng_support@bmc.com}}}
    }, 

</pre>
<p>Job definitions. This is where individual jobs are specified and listed. See below for descriptions of each job in the flow.</p>
<p>Flow statements. These define the relationships of the job, both upstream and downstream.</p>
<pre>"flow00": {"Type":"Flow", "Sequence":["jog-gcs-download", 
           "jog-dflow-gcs-to-bq-fraud", "jog-mc-bq-query", 
           "jog-mc-fm-service"]},
"flow01": {"Type":"Flow", "Sequence":["jog-gcs-download", 
           "jog-dflow-gcs-to-bq-transactions", 
           "jog-mc-bq-query", "jog-mc-fm-service"]},
"flow02": {"Type":"Flow", "Sequence":["jog-gcs-download", 
           "jog-dflow-gcs-to-bq-personal", "jog-mc-bq-query", 
           "jog-mc-fm-service"]} 

</pre>
<h2><span class="TextRun SCXW64342890 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW64342890 BCX0" data-ccp-parastyle="heading 2">How does Control-M schedule pipeline workflows?</span></span></h2>
<p><span class="NormalTextRun SCXW96632814 BCX0" data-ccp-parastyle="Normal (Web)">Control-M uses a server-and-agent model for scheduling. The server is the central engine that manages workflow scheduling and submission to agents, which are lightweight workers. In the </span><span class="NormalTextRun SCXW96632814 BCX0" data-ccp-parastyle="Normal (Web)">demo described in this article, the Control-M server and agent are both running on Google Compute Engine VM instances.</span></p>
<p><span class="TextRun SCXW147537122 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW147537122 BCX0" data-ccp-parastyle="Normal (Web)">Workflows are </span><span class="NormalTextRun AdvancedProofingIssueV2Themed SCXW147537122 BCX0" data-ccp-parastyle="Normal (Web)">most commonly launched</span><span class="NormalTextRun SCXW147537122 BCX0" data-ccp-parastyle="Normal (Web)"> in response to various events such as data arrival but may also be executed automatically based on a predefined schedule. Schedules are very flexible and can refer to business calendars; specify different days of the week, month, or quarter; define cyclic execution, which runs workflows intermittently or </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW147537122 BCX0" data-ccp-parastyle="Normal (Web)">every &#8220;n&#8221; hours or minutes</span><span class="NormalTextRun SCXW147537122 BCX0" data-ccp-parastyle="Normal (Web)">; and so on.</span></span></p>
<h2><span class="TextRun SCXW26310625 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW26310625 BCX0" data-ccp-parastyle="heading 2">How does Control-M process the data?</span></span></h2>
<p><span class="TextRun SCXW208891327 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW208891327 BCX0" data-ccp-parastyle="Normal (Web)">Control-M processes data through a sequence of job types—File Transfer, Dataflow, and SLA Management—each mapped to a distinct stage of the pipeline.</span></span></p>
<h3>File Transfer job type</h3>
<p><span class="TextRun SCXW56059889 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW56059889 BCX0" data-ccp-parastyle="Normal (Web)">The first job, called jog-</span><span class="NormalTextRun SpellingErrorV2Themed SCXW56059889 BCX0" data-ccp-parastyle="Normal (Web)">gcs</span><span class="NormalTextRun SCXW56059889 BCX0" data-ccp-parastyle="Normal (Web)">-download, is of type </span><span class="NormalTextRun SpellingErrorV2Themed SCXW56059889 BCX0" data-ccp-parastyle="Normal (Web)">Job:FileTransfer</span><span class="NormalTextRun SCXW56059889 BCX0" data-ccp-parastyle="Normal (Web)">. This job transfers files from a conventional file system described by </span><span class="NormalTextRun SpellingErrorV2Themed SCXW56059889 BCX0" data-ccp-parastyle="Normal (Web)">ConnectionProfileSrc</span><span class="NormalTextRun SCXW56059889 BCX0" data-ccp-parastyle="Normal (Web)"> to Google Cloud Storage described by </span><span class="NormalTextRun SpellingErrorV2Themed SCXW56059889 BCX0" data-ccp-parastyle="Normal (Web)">ConnectionProfileDest</span><span class="NormalTextRun SCXW56059889 BCX0" data-ccp-parastyle="Normal (Web)">.</span></span></p>
<p><span class="TextRun SCXW180593148 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW180593148 BCX0" data-ccp-parastyle="Normal (Web)">The File Transfer job type can watch for data-related events (file watching) as a prerequisite for data transfer, as well as perform pre/post actions such as deletion of the source after a successful transfer, renaming, source and destination comparison, and restart from the point of failure in the event of an interruption. In the example, this job moves several files from a Linux® host and drops them into Google Cloud Storage buckets.</span></span></p>
<pre>"jog-gcs-download" : {"Type" : "Job:FileTransfer",
        "Host" : "ftpagents",
        "ConnectionProfileSrc" : "smprodMFT",
        "ConnectionProfileDest" : "joggcp",
        "S3BucketName" : "prj1968-bmc-data-platform-foundation",
        "Description" : "First data ingest that triggers downstream applications",
        "FileTransfers" : [
          {
            "TransferType" : "Binary",
            "TransferOption" : "SrcToDestFileWatcher",
            "Src" : "/bmc_personal_details.csv",
            "Dest" : "/bmc_personal_details.csv"
          },
          {
            "TransferType" : "Binary",
            "TransferOption" : "SrcToDestFileWatcher",
            "Src" : "/bmc_fraud_details.csv",
            "Dest" : "/bmc_fraud_details.csv"
          },
          {
            "TransferType" : "Binary",
            "TransferOption" : "SrcToDestFileWatcher",
            "Src" : "/bmc_transaction_details.csv",
            "Dest" : "/bmc_transaction_details.csv"
          } 
        ]
      }, 

</pre>
<h3>Dataflow</h3>
<p>Dataflow jobs are executed to push the newly arrived data into BigQuery. The jobs appear complex, but Google Cloud provides an easy-to-use process to make the definitions simple.</p>
<p><span class="TextRun SCXW233989491 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW233989491 BCX0" data-ccp-parastyle="Normal (Web)">Go to the Dataflow Jobs page (Figure 2). If you have an existing job, choose to Clone it or Create Job from Template. Once </span><span class="NormalTextRun SCXW233989491 BCX0" data-ccp-parastyle="Normal (Web)">you&#8217;ve</span><span class="NormalTextRun SCXW233989491 BCX0" data-ccp-parastyle="Normal (Web)"> provided the desired parameters, click on Equivalent REST at the bottom to get this information (Figure 3), which you can cut and paste directly into the </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW233989491 BCX0" data-ccp-parastyle="Normal (Web)">job&#8217;s</span><span class="NormalTextRun SCXW233989491 BCX0" data-ccp-parastyle="Normal (Web)"> Parameters section.</span></span></p>
<div id="attachment_52270" style="width: 212px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" aria-describedby="caption-attachment-52270" class="wp-image-52270 size-full" src="https://s7280.pcdn.co/wp-content/uploads/2022/09/dataflow-jobs-page.png" alt="Dataflow Jobs page" width="202" height="321" srcset="https://s7280.pcdn.co/wp-content/uploads/2022/09/dataflow-jobs-page.png 202w, https://s7280.pcdn.co/wp-content/uploads/2022/09/dataflow-jobs-page-189x300.png 189w, https://s7280.pcdn.co/wp-content/uploads/2022/09/dataflow-jobs-page-15x24.png 15w, https://s7280.pcdn.co/wp-content/uploads/2022/09/dataflow-jobs-page-23x36.png 23w, https://s7280.pcdn.co/wp-content/uploads/2022/09/dataflow-jobs-page-30x48.png 30w" sizes="auto, (max-width: 202px) 100vw, 202px" /><p id="caption-attachment-52270" class="wp-caption-text">Figure 2: Dataflow Jobs page</p></div>
<div id="attachment_52269" style="width: 350px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" aria-describedby="caption-attachment-52269" class="wp-image-52269 size-full" src="https://s7280.pcdn.co/wp-content/uploads/2022/09/job-parameters-section.png" alt="job Parameters section" width="340" height="314" srcset="https://s7280.pcdn.co/wp-content/uploads/2022/09/job-parameters-section.png 340w, https://s7280.pcdn.co/wp-content/uploads/2022/09/job-parameters-section-300x277.png 300w, https://s7280.pcdn.co/wp-content/uploads/2022/09/job-parameters-section-24x22.png 24w, https://s7280.pcdn.co/wp-content/uploads/2022/09/job-parameters-section-36x33.png 36w, https://s7280.pcdn.co/wp-content/uploads/2022/09/job-parameters-section-48x44.png 48w" sizes="auto, (max-width: 340px) 100vw, 340px" /><p id="caption-attachment-52269" class="wp-caption-text">Figure 3: Cut and paste into job Parameters section</p></div>
<pre>"jog-dflow-gcs-to-bq-fraud": {"Type": "Job:ApplicationIntegrator:AI Google DataFlow",
        "AI-Location": "us-central1",
        "AI-Parameters (JSON Format)": "{"jobName": "jog-dflow-gcs-to-bq-fraud",
        "environment": {        "bypassTempDirValidation": false,
        "tempLocation": "gs://prj1968-bmc-data-platform-foundation/bmc_fraud_details/temp",
        "ipConfiguration": "WORKER_IP_UNSPECIFIED",
        "additionalExperiments": []    },    
        "parameters": {
        "javascriptTextTransformGcsPath": "gs://prj1968-bmc-data-platform-foundation/bmc_fraud_details/bmc_fraud_details_transform.js", 
        "JSONPath": "gs://prj1968-bmc-data-platform-foundation/bmc_fraud_details/bmc_fraud_details_schema.json",
        "javascriptTextTransformFunctionName": "transform",
        "outputTable": "sso-gcp-dba-ctm4-pub-cc10274:bmc_dataplatform_foundation.bmc_fraud_details_V2",
        "inputFilePattern": "gs://prj1968-bmc-data-platform-foundation/bmc_fraud_details/bmc_fraud_details.csv", 
        "bigQueryLoadingTemporaryDirectory": "gs://prj1968-bmc-data-platform-foundation/bmc_fraud_details/tmpbq"    }}",
        "AI-Log Level": "INFO",
        "AI-Template Location (gs://)": "gs://dataflow-templates-us-central1/latest/GCS_Text_to_BigQuery",
        "AI-Project ID": "sso-gcp-dba-ctm4-pub-cc10274",
        "AI-Template Type": "Classic Template",
        "ConnectionProfile": "JOG-DFLOW-MIDENTITY",
        "Host": "gcpagents"
      }, 

</pre>
<h3>SLA management</h3>
<p>This job defines the SLA completion criteria and instructs Control-M to monitor the entire workflow as a single business entity.</p>
<pre>"jog-mc-fm-service": {"Type": "Job:SLAManagement",
	 "ServiceName": "Model testing and scoring for fraud detection",
	 "ServicePriority": "3",
	 "JobRunsDeviationsTolerance": "3",
	 "CompleteIn": {
	    "Time": "20:00"
	  }
	},
</pre>
<p><span class="TextRun SCXW42285845 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW42285845 BCX0" data-ccp-parastyle="Normal (Web)">The </span><span class="NormalTextRun SpellingErrorV2Themed SCXW42285845 BCX0" data-ccp-parastyle="Normal (Web)">ServiceName</span><span class="NormalTextRun SCXW42285845 BCX0" data-ccp-parastyle="Normal (Web)"> specifies a business-relevant name that will appear in notifications or service incidents, as well as in displays for non-technical users, to make it clear which business service may be </span><span class="NormalTextRun SCXW42285845 BCX0" data-ccp-parastyle="Normal (Web)">impacted</span><span class="NormalTextRun SCXW42285845 BCX0" data-ccp-parastyle="Normal (Web)">. Importantly, Control-M uses statistics collected from </span><span class="NormalTextRun SCXW42285845 BCX0" data-ccp-parastyle="Normal (Web)">previous</span><span class="NormalTextRun SCXW42285845 BCX0" data-ccp-parastyle="Normal (Web)"> executions to automatically compute the expected completion so that any deviation can be detected and reported at the earliest possible moment. This gives monitoring teams the maximum opportunity to course-correct before any impact </span><span class="NormalTextRun SCXW42285845 BCX0" data-ccp-parastyle="Normal (Web)">to</span><span class="NormalTextRun SCXW42285845 BCX0" data-ccp-parastyle="Normal (Web)"> business services is detected.</span></span></p>
<h2><span class="TextRun SCXW123172440 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW123172440 BCX0" data-ccp-parastyle="heading 2">How do you </span><span class="NormalTextRun SCXW123172440 BCX0" data-ccp-parastyle="heading 2">monitor</span><span class="NormalTextRun SCXW123172440 BCX0" data-ccp-parastyle="heading 2"> the pipeline state in Control-M?</span></span></h2>
<p><span class="TextRun SCXW184137364 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW184137364 BCX0" data-ccp-parastyle="Normal (Web)">Control-M provides a real-time monitoring interface that shows the status of every job in the pipeline, making it straightforward to </span><span class="NormalTextRun SCXW184137364 BCX0" data-ccp-parastyle="Normal (Web)">identify</span><span class="NormalTextRun SCXW184137364 BCX0" data-ccp-parastyle="Normal (Web)"> failures and </span><span class="NormalTextRun AdvancedProofingIssueV2Themed SCXW184137364 BCX0" data-ccp-parastyle="Normal (Web)">take action</span><span class="NormalTextRun SCXW184137364 BCX0" data-ccp-parastyle="Normal (Web)"> without switching between tools.</span></span></p>
<p><span class="TextRun SCXW101488452 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW101488452 BCX0" data-ccp-parastyle="Normal (Web)">Control-M provides a user interface for monitoring workflows (Figure 4). In the screenshot below, the first job completed successfully and is </span><span class="NormalTextRun SCXW101488452 BCX0" data-ccp-parastyle="Normal (Web)">green,</span><span class="NormalTextRun SCXW101488452 BCX0" data-ccp-parastyle="Normal (Web)"> the next three jobs are executing and depicted in yellow. Jobs that are waiting to run are shown in gray.</span></span></p>
<div id="attachment_52271" style="width: 638px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" aria-describedby="caption-attachment-52271" class="size-full wp-image-52271" src="https://s7280.pcdn.co/wp-content/uploads/2022/09/control-m-monitoring-domain.png" alt="Control-M Monitoring Domain" width="628" height="355" srcset="https://s7280.pcdn.co/wp-content/uploads/2022/09/control-m-monitoring-domain.png 628w, https://s7280.pcdn.co/wp-content/uploads/2022/09/control-m-monitoring-domain-300x170.png 300w, https://s7280.pcdn.co/wp-content/uploads/2022/09/control-m-monitoring-domain-24x14.png 24w, https://s7280.pcdn.co/wp-content/uploads/2022/09/control-m-monitoring-domain-36x20.png 36w, https://s7280.pcdn.co/wp-content/uploads/2022/09/control-m-monitoring-domain-48x27.png 48w" sizes="auto, (max-width: 628px) 100vw, 628px" /><p id="caption-attachment-52271" class="wp-caption-text">Figure 4: Control-M Monitoring Domain</p></div>
<p><span class="TextRun SCXW121837929 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW121837929 BCX0" data-ccp-parastyle="Normal (Web)">You can access the output and logs of every job from the pane on the right-hand side. This capability is vital during daily operations. To </span><span class="NormalTextRun SCXW121837929 BCX0" data-ccp-parastyle="Normal (Web)">monitor</span><span class="NormalTextRun SCXW121837929 BCX0" data-ccp-parastyle="Normal (Web)"> those operations more easily, Control-M provides a single pane to view the output of jobs running on disparate systems without having to connect to each </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW121837929 BCX0" data-ccp-parastyle="Normal (Web)">application&#8217;s</span><span class="NormalTextRun SCXW121837929 BCX0" data-ccp-parastyle="Normal (Web)"> console.</span></span></p>
<p><span class="TextRun SCXW122522281 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW122522281 BCX0" data-ccp-parastyle="Normal (Web)">Control-M also allows you to perform several actions on the jobs in the pipeline, such as hold, rerun, and kill. You sometimes need to perform these actions when troubleshooting a failure or skipping a job, for example.</span></span></p>
<p><span class="TextRun SCXW238565254 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun AdvancedProofingIssueV2Themed SCXW238565254 BCX0" data-ccp-parastyle="Normal (Web)">All of</span><span class="NormalTextRun SCXW238565254 BCX0" data-ccp-parastyle="Normal (Web)"> the functions discussed here are also available from a REST-based API or a CLI.</span></span></p>
<h2>Conclusion</h2>
<p><span class="TextRun SCXW195466094 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW195466094 BCX0" data-ccp-parastyle="Normal (Web)">Coordinating and monitoring workflows across an ML pipeline </span><span class="NormalTextRun SCXW195466094 BCX0" data-ccp-parastyle="Normal (Web)">remains</span><span class="NormalTextRun SCXW195466094 BCX0" data-ccp-parastyle="Normal (Web)"> a complex task, even with the rich set of ML tools that Google Cloud provides. Anytime you need to orchestrate a data pipeline on Google Cloud that combines file transfers, applications, data sources, or infrastructure, Control-M can simplify your workflow orchestration. Control-M integrates, automates, and orchestrates application workflows whether on-premises, on Google Cloud, or in a hybrid environment.</span></span></p>
<h2 aria-level="2"><span data-contrast="none">Frequently asked questions</span><span data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;335559738&quot;:160,&quot;335559739&quot;:80}"> </span></h2>
<p><strong>What is the best tool to orchestrate a data pipeline on Google Cloud?</strong></p>
<p><span data-contrast="auto">Control-M from BMC is purpose-built for orchestrating data pipelines across cloud and hybrid environments. On Google Cloud, Control-M integrates natively with Cloud Storage, Dataflow, BigQuery, and Pub/Sub, enabling teams to define, schedule, and monitor workflows from a single interface using either a graphical editor or RESTful APIs.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What is the difference between Dataflow and Control-M?</strong></p>
<p><span data-contrast="auto">Google Cloud Dataflow is a managed data integration service that moves and transforms data between sources and destinations such as BigQuery. Control-M is a workflow orchestration engine that coordinates and monitors the execution of multiple tools—including Dataflow jobs—as part of a broader end-to-end pipeline. Dataflow handles the data movement; Control-M manages the sequencing, scheduling, and error handling of the entire workflow.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>Can Control-M run jobs as code on Google Cloud?</strong></p>
<p><span data-contrast="auto">Yes. Control-M supports a jobs-as-code approach using JSON and RESTful APIs, which can be integrated into a CI/CD toolchain. Google Cloud developer tools including Cloud Build and Cloud Deploy are compatible with this method.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What Google Cloud services are used in a Control-M data pipeline?</strong></p>
<p><span data-contrast="auto">A typical Control-M-orchestrated pipeline on Google Cloud uses Cloud Storage as the data landing zone, Pub/Sub for streaming data ingestion, Dataflow for ETL processing, BigQuery for analytics and querying, and Data Studio for visualization. Control-M coordinates the sequencing and monitoring of all these services.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>How does Control-M handle pipeline failures on Google Cloud? </strong></p>
<p><span data-contrast="auto">Control-M includes built-in error handling and SLA management capabilities. When a job fails, Control-M can automatically restart it or escalate to a human based on predefined criteria. The SLA Management job type monitors the entire workflow as a single business entity and uses historical execution data to predict completion times, alerting monitoring teams to deviations before business SLAs are breached.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><em>The views and opinions expressed in this post are those of the author and do not necessarily reflect the official position of BMC.</em></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Unlock Your Data Initiatives with DataOps</title>
		<link>https://blogs.bmc.com/unlock-data-initiatives-with-dataops/</link>
		
		<dc:creator><![CDATA[Basil Faruqui]]></dc:creator>
		<pubDate>Mon, 30 Mar 2026 15:55:20 +0000</pubDate>
				<category><![CDATA[Workload Automation Blog]]></category>
		<guid isPermaLink="false">https://blogs.bmc.com/?p=53593</guid>

					<description><![CDATA[<img width="700" height="400" src="https://s7280.pcdn.co/wp-content/uploads/2020/12/BigData_BMC-700x400-1.png" class="attachment-large size-large wp-post-image" alt="BigData" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2020/12/BigData_BMC-700x400-1.png 700w, https://s7280.pcdn.co/wp-content/uploads/2020/12/BigData_BMC-700x400-1-300x171.png 300w, https://s7280.pcdn.co/wp-content/uploads/2020/12/BigData_BMC-700x400-1-24x14.png 24w, https://s7280.pcdn.co/wp-content/uploads/2020/12/BigData_BMC-700x400-1-36x21.png 36w, https://s7280.pcdn.co/wp-content/uploads/2020/12/BigData_BMC-700x400-1-48x27.png 48w" sizes="auto, (max-width: 700px) 100vw, 700px" />DataOps applies agile engineering and DevOps best practices to data management, helping organizations rapidly turn raw data into fully operationalized production deliverables that unlock real business value. For companies struggling to extract results from their data investments, DataOps provides the framework—and the right workflow orchestration platform provides the engine—to run data pipelines reliably at enterprise scale.  Across every industry, companies […]]]></description>
										<content:encoded><![CDATA[<img width="700" height="400" src="https://s7280.pcdn.co/wp-content/uploads/2020/12/BigData_BMC-700x400-1.png" class="attachment-large size-large wp-post-image" alt="BigData" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2020/12/BigData_BMC-700x400-1.png 700w, https://s7280.pcdn.co/wp-content/uploads/2020/12/BigData_BMC-700x400-1-300x171.png 300w, https://s7280.pcdn.co/wp-content/uploads/2020/12/BigData_BMC-700x400-1-24x14.png 24w, https://s7280.pcdn.co/wp-content/uploads/2020/12/BigData_BMC-700x400-1-36x21.png 36w, https://s7280.pcdn.co/wp-content/uploads/2020/12/BigData_BMC-700x400-1-48x27.png 48w" sizes="auto, (max-width: 700px) 100vw, 700px" /><p><span data-contrast="auto">DataOps applies agile engineering and DevOps best practices to data management, helping organizations rapidly turn raw data into fully operationalized production deliverables that unlock real business value. For companies struggling to extract results from their data investments, DataOps provides the framework—and the right workflow orchestration platform provides the engine—to run data pipelines reliably at enterprise scale.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Across every industry, companies continue to put increased focus on gathering data and finding innovative ways to garner actionable insights. Organizations are willing to invest significant time and money to make that happen.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">However, despite high levels of investment, data projects can often yield lackluster results. A recent survey of advanced major analytics programs by </span><a href="https://www.mckinsey.com/capabilities/mckinsey-digital/our-insights/tech-forward/how-companies-can-use-dataops-to-jump-start-advanced-analytics"><span data-contrast="none">McKinsey</span></a><span data-contrast="auto"> found that companies spend 80 percent of their time doing repetitive tasks such as preparing data, where limited value-added work occurs. Additionally, they found that only 10 percent of companies feel they have this issue under control.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">So why are data project failure rates so high despite increased investment and focus?</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Many variables can impact project success. Often cited factors include project complexity and limited talent pools. Data scientists, cloud architects, and data engineers are in short supply globally. Companies are also recognizing that many of their data projects are failing because they struggle to operationalize the data initiatives at scale in production.</span></p>
<h2><span class="TextRun SCXW24572370 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW24572370 BCX0" data-ccp-parastyle="heading 2">How does </span><span class="NormalTextRun SpellingErrorV2Themed SCXW24572370 BCX0" data-ccp-parastyle="heading 2">DataOps</span><span class="NormalTextRun SCXW24572370 BCX0" data-ccp-parastyle="heading 2"> help unlock data initiatives?</span></span></h2>
<p><span data-contrast="auto">DataOps is the application of agile engineering and DevOps best practices to the field of data management—helping organizations rapidly turn new insights into fully operationalized production deliverables that unlock business value from data. By treating data pipelines with the same discipline applied to software delivery, DataOps reduces the time between a raw insight and a production-ready, business-usable output.</span></p>
<p><span data-contrast="auto">The number of organizations adopting DataOps practices to help them unlock their data is increasing exponentially, so much so that analyst firms have started tracking DataOps tools as a market.</span></p>
<p><span class="TextRun SCXW200623172 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW200623172 BCX0" data-ccp-parastyle="Normal (Web)">In 2022, industry analyst Gartner® published the </span></span><a class="Hyperlink SCXW200623172 BCX0" href="https://www.gartner.com/doc/reprints?id=1-2BX1SAUV&amp;ct=221206&amp;st=sb" target="_blank" rel="noreferrer noopener"><span class="TextRun Underlined SCXW200623172 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW200623172 BCX0" data-ccp-charstyle="Hyperlink">Market Guide for DataOps Tools</span></span></a><span class="TextRun SCXW200623172 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW200623172 BCX0" data-ccp-parastyle="Normal (Web)">, in which it provided this market definition:</span></span><span class="EOP Selected SCXW200623172 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span class="TextRun SCXW84160840 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW84160840 BCX0" data-ccp-parastyle="Normal (Web)">&#8220;</span><span class="NormalTextRun SpellingErrorV2Themed SCXW84160840 BCX0" data-ccp-parastyle="Normal (Web)">DataOps</span><span class="NormalTextRun SCXW84160840 BCX0" data-ccp-parastyle="Normal (Web)"> tools provide greater automation and agility over the full life cycle management of data pipelines in order to streamline data operations. The core capabilities of a </span><span class="NormalTextRun SpellingErrorV2Themed SCXW84160840 BCX0" data-ccp-parastyle="Normal (Web)">DataOps</span><span class="NormalTextRun SCXW84160840 BCX0" data-ccp-parastyle="Normal (Web)"> tool include:</span></span></p>
<ul>
<li><strong>Orchestration:</strong> Connectivity, workflow automation, lineage, scheduling, logging, troubleshooting, and alerting</li>
<li><strong>Observability:</strong> Monitoring live/historic workflows, insights into workflow performance and cost metrics, impact analysis</li>
<li><strong>Environment Management:</strong> Infrastructure as code, resource provisioning, environment repository templates, credentials management</li>
<li><strong>Deployment Automation:</strong> Version control, release pipelines, approvals, rollback, and recovery</li>
<li><strong>Test Automation:</strong> Business rules validation, test scripts management, test data management”</li>
</ul>
<p><span class="TextRun SCXW130972338 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW130972338 BCX0" data-ccp-parastyle="Normal (Web)">As the Gartner market definition </span><span class="NormalTextRun SCXW130972338 BCX0" data-ccp-parastyle="Normal (Web)">indicates</span><span class="NormalTextRun SCXW130972338 BCX0" data-ccp-parastyle="Normal (Web)">, orchestration of data pipelines is a key element of </span><span class="NormalTextRun SpellingErrorV2Themed SCXW130972338 BCX0" data-ccp-parastyle="Normal (Web)">DataOps</span><span class="NormalTextRun SCXW130972338 BCX0" data-ccp-parastyle="Normal (Web)"> capabilities. However, data workflow orchestration comes with its own set of challenges.</span></span></p>
<h2><span class="TextRun SCXW16238254 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW16238254 BCX0" data-ccp-parastyle="heading 2">What are the data orchestration challenges?</span></span></h2>
<p><span class="TextRun SCXW256495327 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW256495327 BCX0" data-ccp-parastyle="Normal (Web)">Most data pipeline workflows are immensely complex and run across many disparate applications, data sources, and infrastructure technologies that need to work together. While the goal is to automate these processes in production, the reality is that without a powerful workflow orchestration platform, delivering these projects at enterprise scale can be expensive and often requires </span><span class="NormalTextRun SCXW256495327 BCX0" data-ccp-parastyle="Normal (Web)">significant time</span><span class="NormalTextRun SCXW256495327 BCX0" data-ccp-parastyle="Normal (Web)"> spent doing manual work.</span></span></p>
<p><span class="TextRun SCXW161884825 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW161884825 BCX0" data-ccp-parastyle="Normal (Web)">Data workflow orchestration projects have four key stages: ingestion, storage, processing, and delivering insights to make faster and smarter decisions.</span></span></p>
<div id="attachment_53594" style="width: 634px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" aria-describedby="caption-attachment-53594" class="wp-image-53594 size-full" src="https://s7280.pcdn.co/wp-content/uploads/2024/05/Data-projects-have-four-stages-with-many-moving-parts-across-multiple-technologies.png" alt="Data-projects-have-four-stages-with-many-moving-parts-across-multiple-technologies" width="624" height="265" srcset="https://s7280.pcdn.co/wp-content/uploads/2024/05/Data-projects-have-four-stages-with-many-moving-parts-across-multiple-technologies.png 624w, https://s7280.pcdn.co/wp-content/uploads/2024/05/Data-projects-have-four-stages-with-many-moving-parts-across-multiple-technologies-300x127.png 300w, https://s7280.pcdn.co/wp-content/uploads/2024/05/Data-projects-have-four-stages-with-many-moving-parts-across-multiple-technologies-24x10.png 24w, https://s7280.pcdn.co/wp-content/uploads/2024/05/Data-projects-have-four-stages-with-many-moving-parts-across-multiple-technologies-36x15.png 36w, https://s7280.pcdn.co/wp-content/uploads/2024/05/Data-projects-have-four-stages-with-many-moving-parts-across-multiple-technologies-48x20.png 48w" sizes="auto, (max-width: 624px) 100vw, 624px" /><p id="caption-attachment-53594" class="wp-caption-text">Figure 1. Data projects have four stages with many moving parts across multiple technologies.</p></div>
<p>Ingestion involves collecting data from traditional sources like enterprise resource planning (ERP) and customer resource management (CRM) solutions, financial systems, and many other systems of record in addition to data from modern sources like devices, Internet of Things (IoT) sensors, and social media.</p>
<p>Storage increases the complexity with numerous different tools and technologies that are part of the data pipeline. Where and how you store data depends a lot on persistence, the relative value of the data sets, the refresh rate of your analytics models, and the speed at which you can move the data to processing.</p>
<p>Processing has many of the same challenges. How much pure processing is needed? Is it constant or variable? Is it scheduled, event-driven, or ad hoc? How do you minimize costs? The list goes on and on.</p>
<p>Delivering insights requires moving the data output to analytics systems. This layer is also complex, with a growing number of tools representing the last mile in the data pipeline.</p>
<p>With new data and cloud technologies being frequently introduced, companies are constantly reevaluating their tech stacks. This evolving innovation creates pressure and churn that can be challenging because companies need to easily adopt new technologies and scale them in production. Ultimately, if a new data analytics service is not in production at scale, companies are not getting actionable insights or achieving value.</p>
<h2><span class="TextRun SCXW41691176 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW41691176 BCX0" data-ccp-parastyle="heading 2">What capabilities should a workflow orchestration platform have?</span></span></h2>
<p><span class="NormalTextRun SCXW226783022 BCX0" data-ccp-parastyle="Normal (Web)">Successfully running business-critical workflows at scale in production </span><span class="NormalTextRun SCXW226783022 BCX0" data-ccp-parastyle="Normal (Web)">doesn&#8217;t</span><span class="NormalTextRun SCXW226783022 BCX0" data-ccp-parastyle="Normal (Web)"> happen by accident. The right workflow orchestration platform can help you streamline your data pipelines and get the actionable insights you need.</span></p>
<p><span class="TextRun SCXW171471750 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW171471750 BCX0" data-ccp-parastyle="Normal (Web)">Here are eight essential capabilities to look for in your workflow orchestration platform:</span></span></p>
<ol>
<li>Support heterogeneous workflows: Companies are rapidly moving to the cloud, and for the foreseeable future will have workflows across a highly complex mix of hybrid environments. For many, this will include supporting the mainframe and distributed systems across the data center and multiple private and/or public clouds. If your orchestration platform cannot handle the diversity of applications and underlying infrastructure, you will have a highly fragmented automation strategy with many silos of automation that require cumbersome custom integrations to handle cross-platform workflow dependencies.</li>
<li>Service <span class="NormalTextRun SCXW123969367 BCX0" data-ccp-parastyle="Normal (Web)">level agreement (SLA) management: Business workflows—ranging from ML models predicting risk to financial close and payment settlements—all have completion SLAs that are sometimes governed by guidelines set by regulatory agencies. Your orchestration platform must be able to understand and </span><span class="NormalTextRun SCXW123969367 BCX0" data-ccp-parastyle="Normal (Web)">notify you</span><span class="NormalTextRun SCXW123969367 BCX0" data-ccp-parastyle="Normal (Web)"> of task failures and delays in complex workflows, and it needs to be able to map issues to broader business impacts.</span></li>
<li>Error handling and notifications: <span class="TextRun SCXW81148069 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW81148069 BCX0" data-ccp-parastyle="Normal (Web)"> When running in production, even the best-designed workflows will have failures and delays. It is vital that the right teams are notified so that lengthy war room discussions just to figure out who needs to work on a problem can be avoided. Your orchestration platform must automatically send notifications to the right teams at the right time.</span></span></li>
<li>Self-healing and remediation: When teams respond to job failures within business workflows, they take corrective action, such as restarting a job, deleting a file, or flushing a cache or temp table. Your orchestration platform should enable automation engineers to configure such actions to happen automatically the next time the same problem occurs.</li>
<li><span class="TextRun SCXW225492777 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW225492777 BCX0" data-ccp-parastyle="Normal (Web)">End-to-end visibility: Workflows execute interconnected business processes across hybrid tech stacks. Your orchestration platform should be able to clearly show the lineage of your workflows. This is integral to helping you understand the relationships between applications and the business processes they support. This is also important for change management—when making changes, it is vital to see what happens upstream and downstream from a process.</span></span><span class="EOP Selected SCXW225492777 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span class="TextRun SCXW85926558 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW85926558 BCX0" data-ccp-parastyle="Normal (Web)">Self-service user experience (UX) for multiple personas: Workflow orchestration is a team sport with many stakeholders such as data teams, developers, operations, business process owners, and more. Each team has different use cases and preferences for how they want to interact with the orchestration tools. This means your orchestration platform must offer the right user interface (UI) and UX for each team so they can </span><span class="NormalTextRun SCXW85926558 BCX0" data-ccp-parastyle="Normal (Web)">benefit</span><span class="NormalTextRun SCXW85926558 BCX0" data-ccp-parastyle="Normal (Web)"> from the technology.</span></span><span class="EOP Selected SCXW85926558 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span class="TextRun SCXW73311739 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW73311739 BCX0" data-ccp-parastyle="Normal (Web)">Production standards: Running workflows in production requires adherence to standards, which means using correct naming conventions, error-handling patterns, and so on. Your orchestration platform should have a mechanism that </span><span class="NormalTextRun SCXW73311739 BCX0" data-ccp-parastyle="Normal (Web)">provides</span><span class="NormalTextRun SCXW73311739 BCX0" data-ccp-parastyle="Normal (Web)"> </span><span class="NormalTextRun SCXW73311739 BCX0" data-ccp-parastyle="Normal (Web)">a very simple</span><span class="NormalTextRun SCXW73311739 BCX0" data-ccp-parastyle="Normal (Web)"> way to define such standards and guide users to the </span><span class="NormalTextRun SCXW73311739 BCX0" data-ccp-parastyle="Normal (Web)">appropriate standards</span><span class="NormalTextRun SCXW73311739 BCX0" data-ccp-parastyle="Normal (Web)"> when they are building workflows.</span></span></li>
<li><span class="TextRun SCXW170865653 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW170865653 BCX0" data-ccp-parastyle="Normal (Web)">Support DevOps practices: As companies adopt DevOps practices such as continuous integration and continuous deployment (CI/CD) pipelines, the workflow development, modification, and even infrastructure deployment of workflows, your orchestration platform should be able to fit into modern release practices.</span></span><span class="EOP Selected SCXW170865653 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
</ol>
<h2><span class="TextRun SCXW261052232 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW261052232 BCX0" data-ccp-parastyle="heading 2">How do Control-M and Control-M SaaS support </span><span class="NormalTextRun SpellingErrorV2Themed SCXW261052232 BCX0" data-ccp-parastyle="heading 2">DataOps</span><span class="NormalTextRun SCXW261052232 BCX0" data-ccp-parastyle="heading 2">?</span></span><span class="EOP Selected SCXW261052232 BCX0" data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;335559738&quot;:160,&quot;335559739&quot;:80}"> </span></h2>
<p><span class="TextRun SCXW251864231 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SpellingErrorV2Themed SCXW251864231 BCX0" data-ccp-parastyle="Normal (Web)">DataOps</span><span class="NormalTextRun SCXW251864231 BCX0" data-ccp-parastyle="Normal (Web)"> tools and methodologies can help you make the best use of your data investment. But if you want to succeed in your </span><span class="NormalTextRun SpellingErrorV2Themed SCXW251864231 BCX0" data-ccp-parastyle="Normal (Web)">DataOps</span><span class="NormalTextRun SCXW251864231 BCX0" data-ccp-parastyle="Normal (Web)"> journey, you must be able to operationalize the data. Control-M (self-hosted) and Control-M SaaS provide a layer of abstraction to simplify the orchestration of complex data pipelines. These application and data workflow orchestration platforms enable end-to-end visibility and predictive SLAs across any data technology or infrastructure.</span></span></p>
<div id="attachment_53595" style="width: 634px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" aria-describedby="caption-attachment-53595" class="wp-image-53595 size-full" src="https://s7280.pcdn.co/wp-content/uploads/2024/05/Control-M-is-a-layer-of-abstraction-to-simplify-complex-data-pipelines.png" alt="Control-M is a layer of abstraction to simplify complex data pipelines" width="624" height="265" srcset="https://s7280.pcdn.co/wp-content/uploads/2024/05/Control-M-is-a-layer-of-abstraction-to-simplify-complex-data-pipelines.png 624w, https://s7280.pcdn.co/wp-content/uploads/2024/05/Control-M-is-a-layer-of-abstraction-to-simplify-complex-data-pipelines-300x127.png 300w, https://s7280.pcdn.co/wp-content/uploads/2024/05/Control-M-is-a-layer-of-abstraction-to-simplify-complex-data-pipelines-24x10.png 24w, https://s7280.pcdn.co/wp-content/uploads/2024/05/Control-M-is-a-layer-of-abstraction-to-simplify-complex-data-pipelines-36x15.png 36w, https://s7280.pcdn.co/wp-content/uploads/2024/05/Control-M-is-a-layer-of-abstraction-to-simplify-complex-data-pipelines-48x20.png 48w" sizes="auto, (max-width: 624px) 100vw, 624px" /><p id="caption-attachment-53595" class="wp-caption-text">Figure 2. Control-M is a layer of abstraction to simplify complex data pipelines.</p></div>
<p><span class="TextRun SCXW7626780 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW7626780 BCX0" data-ccp-parastyle="Normal (Web)">Control-M and Control-M SaaS can help you orchestrate your data pipelines, put your data to effective use, and improve your data-driven business outcomes. Both platforms are used by thousands of companies globally and are proven to help companies run data pipeline workflows in production at scale.</span></span></p>
<p><span class="TextRun SCXW135403943 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW135403943 BCX0" data-ccp-parastyle="Normal (Web)">Here are some examples of the robust capabilities Control-M and Control-M SaaS have and how they can help you streamline your data pipeline workflow orchestration:</span></span></p>
<h3><span class="TextRun SCXW188285147 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW188285147 BCX0" data-ccp-parastyle="heading 3">Robust integrations</span></span></h3>
<p>The tools required to run a modern business vary widely. Often, each department utilizes its own technologies, requiring manual scripting to connect workflows across the business. Control-M and Control-M SaaS feature a vast <a href="/it-solutions/control-m-integrations.html#&amp;sortCriteria=recommended&amp;category=mp">library of out-of-the-box integrations</a> that allow businesses to orchestrate the latest technologies.</p>
<h3>SLA management and impact analysis</h3>
<p>With Control-M and Control-M SaaS, you can track the status of business service levels along with corresponding workflows, so you know exactly how business services are performing at any given time. The two platforms can predict that a service will be late if a job is delayed or has failed upstream because they are using historical data to calculate how long a downstream job usually takes to run. Using this data, they can notify stakeholders not only that a particular job is late, but which business services are at risk of being delayed.</p>
<h3>Python client</h3>
<p><span class="TextRun SCXW78164153 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW78164153 BCX0" data-ccp-parastyle="Normal (Web)">Many teams within an organization need to interact with your workflow orchestration platform for </span><span class="NormalTextRun SCXW78164153 BCX0" data-ccp-parastyle="Normal (Web)">various reasons</span><span class="NormalTextRun SCXW78164153 BCX0" data-ccp-parastyle="Normal (Web)">. Developers are a particularly important stakeholder in the orchestration process. They develop </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW78164153 BCX0" data-ccp-parastyle="Normal (Web)">the applications</span><span class="NormalTextRun SCXW78164153 BCX0" data-ccp-parastyle="Normal (Web)"> that will run in production and be orchestrated by Control-M and Control-M SaaS. The Python client allows developers to natively invoke their functions from their Python code.</span></span></p>
<h3>Visibility for business users</h3>
<p><span data-contrast="auto">Business users are an important stakeholder, as well. They are ultimately responsible for the timely delivery of the services they own. With the Control-M mobile app and web interface, they can track the status of their workflows anytime, from anywhere, without having to contact the application teams or operations for status updates.</span></p>
<p><span data-contrast="auto">The need for data is on the rise and shows no signs of abating, which means that having the ability to store, process, and operationalize that data will remain crucial to the success of any organization. DataOps practices backed by the powerful data orchestration capabilities of Control-M and Control-M SaaS can help you orchestrate data pipelines, streamline the data delivery process, and improve business outcomes.</span></p>
<p><span class="TextRun SCXW244516046 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW244516046 BCX0" data-ccp-parastyle="Normal (Web)">To learn more about how </span></span><a class="Hyperlink SCXW244516046 BCX0" href="/it-solutions/control-m-big-data.html" target="_blank" rel="noreferrer noopener"><span class="TextRun Underlined SCXW244516046 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW244516046 BCX0" data-ccp-charstyle="Hyperlink">Control-M</span></span></a><span class="TextRun SCXW244516046 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW244516046 BCX0" data-ccp-parastyle="Normal (Web)">/Control-M SaaS can help you deliver data-driven outcomes faster, visit our website.</span></span><span class="EOP Selected SCXW244516046 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<ol>
<li>*<em>Market Guide for DataOps Tools</em>; December 5, 2022; Robert Thanaraj, Sharat Menon, Ankush Jain</li>
</ol>
<p><span class="TextRun SCXW125110687 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW125110687 BCX0" data-ccp-parastyle="Normal (Web)">GARTNER is a registered trademark and service mark of Gartner, Inc. and/or its affiliates in the U.S. and internationally and is used </span><span class="NormalTextRun SCXW125110687 BCX0" data-ccp-parastyle="Normal (Web)">herein</span><span class="NormalTextRun SCXW125110687 BCX0" data-ccp-parastyle="Normal (Web)"> with permission. All rights reserved.</span></span></p>
<h2 aria-level="2"><span data-contrast="none">Frequently asked questions</span><span data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;335559738&quot;:160,&quot;335559739&quot;:80}"> </span></h2>
<p><strong>What is DataOps and why are organizations adopting it?</strong></p>
<p><span data-contrast="auto">DataOps is the application of agile engineering and DevOps best practices to data management—helping organizations rapidly operationalize data insights into production-ready deliverables. Organizations adopt DataOps because the complexity of modern data environments makes it nearly impossible to deliver reliable, scalable data pipelines without a structured framework for automation, orchestration, and governance.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What are the four stages of data workflow orchestration?</strong></p>
<p><span data-contrast="auto">Data workflow orchestration projects move through four stages: ingestion (collecting data from traditional and modern sources), storage (managing persistence, value, and refresh rates), processing (handling compute requirements and scheduling), and delivering insights (routing data output to analytics systems for decision-making).</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What are the most important capabilities in a workflow orchestration platform?</strong></p>
<p><span data-contrast="auto">The eight essential capabilities are: support for heterogeneous workflows across hybrid environments, SLA management and business impact mapping, error handling and automatic notifications, self-healing and remediation, end-to-end workflow visibility and lineage, self-service UX for multiple personas, production standards enforcement, and support for DevOps and CI/CD practices.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>How do Control-M and Control-M SaaS simplify data pipeline orchestration?</strong></p>
<p><span data-contrast="auto">Control-M and Control-M SaaS act as an abstraction layer that simplifies complex data pipeline orchestration by providing a vast library of out-of-the-box integrations, predictive SLA management, dependency-aware scheduling, and end-to-end visibility across any data technology or infrastructure. Both platforms are used by thousands of companies globally to run data pipeline workflows reliably at enterprise scale.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What is the difference between Control-M and Control-M SaaS?</strong></p>
<p><span data-contrast="auto">Control-M is a self-hosted workflow orchestration platform, while Control-M SaaS is a fully managed, cloud-delivered version of the same platform. Both provide core capabilities—including SLA management, predictive impact analysis, robust integrations, and end-to-end visibility—but Control-M SaaS eliminates infrastructure management overhead, making it well-suited for organizations standardizing on cloud-first operations.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><em>The views and opinions expressed in this post are those of the author and do not necessarily reflect the official position of BMC.</em></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Why Orchestration—not More Agents—is the Key to Scaling Enterprise AI</title>
		<link>https://blogs.bmc.com/why-orchestration-not-more-agents-is-the-key-to-scaling-enterprise-ai/</link>
		
		<dc:creator><![CDATA[Basil Faruqui]]></dc:creator>
		<pubDate>Mon, 30 Mar 2026 15:02:00 +0000</pubDate>
				<category><![CDATA[Workload Automation Blog]]></category>
		<guid isPermaLink="false">https://blogs.bmc.com/?p=55856</guid>

					<description><![CDATA[<img width="810" height="405" src="https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-1024x512.png" class="attachment-large size-large wp-post-image" alt="Blue purple gradient charts and screens_1400x700[2]" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-1024x512.png 1024w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-300x150.png 300w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-768x384.png 768w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-810x405.png 810w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-1140x570.png 1140w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-24x12.png 24w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-36x18.png 36w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-48x24.png 48w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002.png 1400w" sizes="auto, (max-width: 810px) 100vw, 810px" />The multi-agent AI era isn’t coming—it’s already here. According to Deloitte, 75% of organizations are investing in AI agents, driving a surge in enterprise adoption. And according to IDC, this isn’t incremental—it’s a structural shift, with agentic AI–driven investment expected to reach $1.3 trillion by 2029. On the surface, more agents should mean more value. […]]]></description>
										<content:encoded><![CDATA[<img width="810" height="405" src="https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-1024x512.png" class="attachment-large size-large wp-post-image" alt="Blue purple gradient charts and screens_1400x700[2]" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-1024x512.png 1024w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-300x150.png 300w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-768x384.png 768w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-810x405.png 810w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-1140x570.png 1140w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-24x12.png 24w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-36x18.png 36w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002-48x24.png 48w, https://s7280.pcdn.co/wp-content/uploads/2022/04/Blue-purple-gradient-charts-and-screens_1400x7002.png 1400w" sizes="auto, (max-width: 810px) 100vw, 810px" /><p>The multi-agent AI era isn’t coming—it’s already here. According to <a href="https://www.deloitte.com/us/en/insights/industry/technology/technology-media-and-telecom-predictions/2026/ai-agent-orchestration.html?id=us:2el:3dp:wsjspon:awa:WSJCMO:2026:WSJFY26">Deloitte</a>, 75% of organizations are investing in AI agents, driving a surge in enterprise adoption. And according to <a href="https://my.idc.com/getdoc.jsp?containerId=prUS53765225">IDC</a>, this isn’t incremental—it’s a structural shift, with agentic AI–driven investment expected to reach $1.3 trillion by 2029.</p>
<p>On the surface, more agents should mean more value.</p>
<p>It doesn’t.</p>
<p>The very force accelerating AI adoption—agent proliferation—may ultimately constrain its impact. As Deloitte notes, once organizations move toward multi‑agent systems, orchestration becomes essential to unlocking their full potential. Yet many enterprises still frame the problem as one of intelligence: bigger models, smarter agents, more autonomy.</p>
<p>That framing is incomplete. The real challenge facing enterprises today isn’t intelligence—it’s execution.</p>
<p>This article explores why orchestration, not additional agents, is the critical missing layer in enterprise AI. It explains how agent sprawl creates complexity, why agent‑only orchestration falls short, and why enterprises must treat orchestration as a control plane—coordinating agents, workflows, data pipelines, and legacy systems—to reliably translate AI into real business outcomes.</p>
<h2>Five realities enterprises must confront</h2>
<p>The challenge isn’t whether agents will be adopted. That’s already happening. The real question is whether enterprises are prepared for what comes next.</p>
<h3>1. Agent sprawl is inevitable</h3>
<p>As agents deliver value, organizations will deploy more of them—quickly, on a multitude of platforms. What starts as a targeted approach becomes a distributed ecosystem of autonomous components. Left unchecked, this fragmentation creates a coordination problem—multiple agents making decisions across disconnected environments with no shared understanding of timing, dependencies, or outcomes. It’s a bit like the “Not Hot Dog” app from the series <a href="https://www.engadget.com/2017-05-15-not-hotdog-app-hbo-silicon-valley.html">Silicon Valley</a>—a model that could perfectly identify a hot dog and confidently label everything else as “not hot dog.” Technically impressive. Practically useless beyond a very narrow context</p>
<h3>2. Orchestrating agents isn’t enough</h3>
<p>The instinctive response is to orchestrate the agents themselves. But that only solves part of the problem.</p>
<p>Agents don’t operate in isolation—they plug into larger business processes. Financial close, trade reconciliation, inventory replenishment, even data pipelines that power inference, RAG, and BI all remain multi-step workflows. Agents may automate decisions within them, but they don’t run them end-to-end.</p>
<p>Which means orchestration can’t be designed around agents alone. It has to coordinate agents alongside scripts, APIs, batch jobs, and serverless functions that make up the rest of the process.</p>
<p>Otherwise, you’re not eliminating complexity—you’re creating another orchestration silo that still has to be connected to everything else.</p>
<h3>3. AI-ready data doesn’t solve itself</h3>
<p>Another emerging lesson is that agents are only as good as the data they consume. As enterprises invest heavily in models and agents, many discover that the real bottleneck is data readiness. Fragmented, outdated, or poorly governed data leads to unreliable outputs. What is new, however, is orchestration’s role in resolving that bottleneck. Preparing AI-ready data requires coordinating data pipelines, application workflows, and event triggers across the enterprise. The intelligence layer depends on that foundation.</p>
<h3>4. The enterprise is more hybrid than ever</h3>
<p>Despite the hype around new technologies, most enterprises operate in deeply hybrid environments. Mainframes remain the lifeblood of many of the world’s largest companies—powering core transactions and systems of record—while cloud-native platforms and microservices drive new digital experiences and AI innovation. Modern data tools interact with both, and critical processes now span generations of infrastructure. These systems aren’t disappearing anytime soon. The challenge isn’t replacing them—it’s ensuring that new AI-driven capabilities work alongside them. That’s where orchestration across the entire stack becomes essential.</p>
<h3>5. Reliability still defines success</h3>
<p>In the race to deploy the newest AI tools, it’s easy to overlook something fundamental: reliability. Enterprise workflow orchestration has long been judged by a simple standard—it just works. Think of it like a Swiss watch: precise, dependable, and trusted to run critical operations. AI systems must meet that same bar. Autonomy is powerful, but enterprises won’t accept fragile automation in mission-critical environments. The orchestration layer must ensure workflows remain predictable, auditable, and resilient—even as intelligence becomes more distributed.</p>
<h2>The path from complexity to simplicity</h2>
<p>Most enterprise problems aren’t glamorous. It’s easy to get excited about frontier models, GPUs, and agents that can reason and act—that’s where the headlines are. But the problems haven’t changed: How do we accelerate financial close? How do we detect and prevent fraud before it impacts customers? How do we execute trades reliably at market speed? How do we keep shelves stocked and orders fulfilled? How do we ensure critical healthcare data is available when it’s needed most?</p>
<p>Simple in nature. Relentless in execution.</p>
<p>Delivering those outcomes means coordinating workflows across agents and traditional systems like ERPs, CRMs, data lakes that span everything from multiple clouds to mainframe systems—all while meeting SLAs, audit, traceability, and explainability requirements. There’s nothing flashy about that. But without it, AI stays in the lab and never graduates to production environments, which is where systems deliver business value.</p>
<p>This is why orchestration isn’t just a tool—it’s a strategy. A control plane for execution.</p>
<p>The goal isn’t more agents—it’s better outcomes. Achieving that requires something often overlooked: simplicity. Orchestration is what turns complexity into simplicity<strong>.</strong> As Leonardo da Vinci put it, <strong>“</strong><em>Simplicity is the ultimate sophistication</em><strong>.”</strong></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>5 Reasons ETL is the Wrong Approach for Mainframe Data Migration</title>
		<link>https://blogs.bmc.com/5-reasons-etl-is-the-wrong-approach-for-mainframe-data-migration/</link>
		
		<dc:creator><![CDATA[Gil Peleg]]></dc:creator>
		<pubDate>Mon, 30 Mar 2026 13:51:33 +0000</pubDate>
				<category><![CDATA[Mainframe Blog]]></category>
		<guid isPermaLink="false">https://blogs.bmc.com/?p=52984</guid>

					<description><![CDATA[<img width="700" height="400" src="https://s7280.pcdn.co/wp-content/uploads/2017/10/CloudComplianceExplained4KeysforSuccess_Final.jpg.optimal.jpg" class="attachment-large size-large wp-post-image" alt="" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2017/10/CloudComplianceExplained4KeysforSuccess_Final.jpg.optimal.jpg 700w, https://s7280.pcdn.co/wp-content/uploads/2017/10/CloudComplianceExplained4KeysforSuccess_Final-300x171.jpg.optimal.jpg 300w, https://s7280.pcdn.co/wp-content/uploads/2017/10/CloudComplianceExplained4KeysforSuccess_Final-24x14.jpg.optimal.jpg 24w, https://s7280.pcdn.co/wp-content/uploads/2017/10/CloudComplianceExplained4KeysforSuccess_Final-36x21.jpg.optimal.jpg 36w, https://s7280.pcdn.co/wp-content/uploads/2017/10/CloudComplianceExplained4KeysforSuccess_Final-48x27.jpg.optimal.jpg 48w" sizes="auto, (max-width: 700px) 100vw, 700px" />ETL (extract, transform, and load) is the wrong approach for mainframe data migration because it was built for structured, database-to-database transfers—not the flexible, cloud-ready data movement that modern mainframe environments require. ETL’s complexity, labor demands, processing costs, and inability to handle unstructured data make it a poor fit for organizations pursuing mainframe data migration today. […]]]></description>
										<content:encoded><![CDATA[<img width="700" height="400" src="https://s7280.pcdn.co/wp-content/uploads/2017/10/CloudComplianceExplained4KeysforSuccess_Final.jpg.optimal.jpg" class="attachment-large size-large wp-post-image" alt="" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2017/10/CloudComplianceExplained4KeysforSuccess_Final.jpg.optimal.jpg 700w, https://s7280.pcdn.co/wp-content/uploads/2017/10/CloudComplianceExplained4KeysforSuccess_Final-300x171.jpg.optimal.jpg 300w, https://s7280.pcdn.co/wp-content/uploads/2017/10/CloudComplianceExplained4KeysforSuccess_Final-24x14.jpg.optimal.jpg 24w, https://s7280.pcdn.co/wp-content/uploads/2017/10/CloudComplianceExplained4KeysforSuccess_Final-36x21.jpg.optimal.jpg 36w, https://s7280.pcdn.co/wp-content/uploads/2017/10/CloudComplianceExplained4KeysforSuccess_Final-48x27.jpg.optimal.jpg 48w" sizes="auto, (max-width: 700px) 100vw, 700px" /><p><span class="TextRun SCXW510919 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW510919 BCX0" data-ccp-parastyle="Normal (Web)">ETL (extract, transform, and load) is the wrong approach for mainframe data migration because it was built for structured, database-to-database transfers—not the flexible, cloud-ready data movement that modern mainframe environments require. ETL&#8217;s complexity, labor demands, processing costs, and inability to handle unstructured data make it a poor fit for organizations pursuing mainframe data migration today. ELT (extract, load, and transform), which moves raw data to its destination first and transforms it there, is better aligned with how mainframe modernization </span><span class="NormalTextRun AdvancedProofingIssueV2Themed SCXW510919 BCX0" data-ccp-parastyle="Normal (Web)">actually works</span><span class="NormalTextRun SCXW510919 BCX0" data-ccp-parastyle="Normal (Web)">.</span></span></p>
<p><span class="TextRun SCXW93665196 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW93665196 BCX0" data-ccp-parastyle="Normal (Web)">Change is good—a familiar mantra, but one </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW93665196 BCX0" data-ccp-parastyle="Normal (Web)">not</span><span class="NormalTextRun SCXW93665196 BCX0" data-ccp-parastyle="Normal (Web)"> always easy to practice. When it comes to moving toward a new way of handling data, mainframe organizations, which have earned their keep by delivering the IT equivalent of corporate-wide insurance policies (rugged, reliable, and risk-averse), naturally look with caution on new concepts like </span></span><a class="Hyperlink SCXW93665196 BCX0" href="/learn/etl-extract-transform-load.html" rel="noreferrer noopener"><span class="TextRun Underlined SCXW93665196 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW93665196 BCX0" data-ccp-charstyle="Hyperlink">extract, load, and transform</span></span></a><span class="TextRun SCXW93665196 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW93665196 BCX0" data-ccp-parastyle="Normal (Web)"> (ELT).</span></span><span class="EOP Selected SCXW93665196 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span class="TextRun SCXW196004912 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW196004912 BCX0" data-ccp-parastyle="Normal (Web)">Positioned as a lighter and faster alternative to more traditional data handling procedures such as extract, transform, and load (ETL), ELT definitely invites scrutiny.</span><span class="NormalTextRun SCXW196004912 BCX0" data-ccp-parastyle="Normal (Web)"> And that scrutiny can be worthwhile.</span></span><span class="EOP Selected SCXW196004912 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span class="TextRun SCXW247899181 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW247899181 BCX0" data-ccp-parastyle="Normal (Web)">Definitions provided by SearchDataManagement.com say that </span></span><a class="Hyperlink SCXW247899181 BCX0" href="https://searchdatamanagement.techtarget.com/definition/Extract-Load-Transform-ELT" target="_blank" rel="noreferrer noopener"><span class="TextRun Underlined SCXW247899181 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW247899181 BCX0" data-ccp-charstyle="Hyperlink">ELT</span></span></a><span class="TextRun SCXW247899181 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW247899181 BCX0" data-ccp-parastyle="Normal (Web)"> is &#8220;a data integration process for transferring raw data from a source server to a data system (such as a data warehouse or data lake) on a target server and then preparing the information for downstream uses.&#8221; In contrast, another source defines </span></span><a class="Hyperlink SCXW247899181 BCX0" href="https://www.webopedia.com/definitions/etl/" target="_blank" rel="noreferrer noopener"><span class="TextRun Underlined SCXW247899181 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW247899181 BCX0" data-ccp-charstyle="Hyperlink">ETL</span></span></a><span class="TextRun SCXW247899181 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW247899181 BCX0" data-ccp-parastyle="Normal (Web)"> as &#8220;three database functions that are combined into one tool to pull data out of one database and place it into another database.&#8221;</span></span><span class="EOP Selected SCXW247899181 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span class="TextRun SCXW239333490 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW239333490 BCX0" data-ccp-parastyle="Normal (Web)">The crucial functional difference is that ETL focuses exclusively on database-to-database transfer, while ELT is open-ended and flexible. In the mainframe world, ETL is a tool with a more limited focus—ELT is focused on jump-starting the future.</span></span></p>
<h2><span class="TextRun SCXW180715764 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW180715764 BCX0" data-ccp-parastyle="heading 2">Why does ETL fail for mainframe data migration?</span></span></h2>
<p><span class="TextRun SCXW147363577 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW147363577 BCX0" data-ccp-parastyle="Normal (Web)">ETL falls short across five key dimensions: complexity, labor intensity, processing bottlenecks, structural rigidity, and high processing costs. Here is a closer look at </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW147363577 BCX0" data-ccp-parastyle="Normal (Web)">each</span><span class="NormalTextRun SCXW147363577 BCX0" data-ccp-parastyle="Normal (Web)">.</span></span></p>
<h3><span class="TextRun SCXW85504748 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW85504748 BCX0" data-ccp-parastyle="heading 3">1. ETL is too complex</span></span></h3>
<p><span class="TextRun SCXW239437273 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW239437273 BCX0" data-ccp-parastyle="Normal (Web)">ETL was not originally designed to handle all the tasks it is now being asked to do. In the early days, ETL was often applied to pull data from one relational structure and fit it into a different relational structure—including cleansing the data along the way.</span></span><span class="EOP Selected SCXW239437273 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span class="TextRun SCXW207948146 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW207948146 BCX0" data-ccp-parastyle="Normal (Web)">For example, a traditional relational database management system (RDBMS) </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW207948146 BCX0" data-ccp-parastyle="Normal (Web)">can get</span><span class="NormalTextRun SCXW207948146 BCX0" data-ccp-parastyle="Normal (Web)"> befuddled by numeric data where it is expecting alpha data, or by the presence of obsolete address abbreviations. ETL is </span><span class="NormalTextRun SCXW207948146 BCX0" data-ccp-parastyle="Normal (Web)">optimized</span><span class="NormalTextRun SCXW207948146 BCX0" data-ccp-parastyle="Normal (Web)"> for that kind of painstaking, field-by-field data checking, &#8220;cleaning,&#8221; and movement—but not for feeding a Hadoop database or modern data lake. ETL was not invented to take advantage of all the ways data originates and all the ways it can be used today.</span></span><span class="EOP Selected SCXW207948146 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3>2.<span class="TextRun SCXW231091511 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW231091511 BCX0" data-ccp-parastyle="heading 3">ETL is labor-intensive</span></span></h3>
<p><span class="TextRun SCXW20382600 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW20382600 BCX0" data-ccp-parastyle="Normal (Web)">Because the </span></span><a class="Hyperlink SCXW20382600 BCX0" href="/learn/reverse-etl.html" rel="noreferrer noopener"><span class="TextRun Underlined SCXW20382600 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW20382600 BCX0" data-ccp-charstyle="Hyperlink">ETL</span></span></a><span class="TextRun SCXW20382600 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW20382600 BCX0" data-ccp-parastyle="Normal (Web)"> process is built around transformation, everything depends on the </span><span class="NormalTextRun SCXW20382600 BCX0" data-ccp-parastyle="Normal (Web)">timely</span><span class="NormalTextRun SCXW20382600 BCX0" data-ccp-parastyle="Normal (Web)"> completion of that transformation step. With larger amounts of data in play—think Big Data—</span><span class="NormalTextRun SCXW20382600 BCX0" data-ccp-parastyle="Normal (Web)">the required transformation times can become inconvenient or impractical, turning ETL into a functional and computational bottleneck.</span></span><span class="EOP Selected SCXW20382600 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3><span class="TextRun SCXW74165533 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW74165533 BCX0" data-ccp-parastyle="heading 3">3.ETL demands structure</span></span></h3>
<p><span class="TextRun SCXW111661539 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW111661539 BCX0" data-ccp-parastyle="Normal (Web)">Because the </span></span><a class="Hyperlink SCXW111661539 BCX0" href="/learn/reverse-etl.html" rel="noreferrer noopener"><span class="TextRun Underlined SCXW111661539 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW111661539 BCX0" data-ccp-charstyle="Hyperlink">ETL</span></span></a><span class="TextRun SCXW111661539 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW111661539 BCX0" data-ccp-parastyle="Normal (Web)"> process is built around transformation, everything depends on the </span><span class="NormalTextRun SCXW111661539 BCX0" data-ccp-parastyle="Normal (Web)">timely</span><span class="NormalTextRun SCXW111661539 BCX0" data-ccp-parastyle="Normal (Web)"> completion of that transformation step. With larger amounts of data in play—think Big Data—</span><span class="NormalTextRun SCXW111661539 BCX0" data-ccp-parastyle="Normal (Web)">the required transformation times can become inconvenient or impractical, turning ETL into a functional and computational bottleneck.</span></span><span class="EOP Selected SCXW111661539 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3><span class="TextRun SCXW91435748 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW91435748 BCX0" data-ccp-parastyle="heading 3">4.ETL demands structure</span></span></h3>
<p><span class="TextRun SCXW221214150 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW221214150 BCX0" data-ccp-parastyle="Normal (Web)">ETL is not designed for unstructured data and can add complexity rather than value when asked to handle it. ETL works best for traditional databases but does not help much with the massive waves of unstructured data that organizations need to process today.</span></span></p>
<h3>5.<span class="TextRun SCXW216072672 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW216072672 BCX0" data-ccp-parastyle="heading 3">ETL has high processing costs</span></span></h3>
<p><span class="TextRun SCXW191537361 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW191537361 BCX0" data-ccp-parastyle="Normal (Web)">ETL is especially challenging for mainframes because mainframe workloads </span><span class="NormalTextRun SCXW191537361 BCX0" data-ccp-parastyle="Normal (Web)">generally incur</span><span class="NormalTextRun SCXW191537361 BCX0" data-ccp-parastyle="Normal (Web)"> MSU (million service unit) processing charges—burdening systems that need to be handling real-time workloads at the same time. ELT, by contrast, can be </span><span class="NormalTextRun SCXW191537361 BCX0" data-ccp-parastyle="Normal (Web)">accomplished</span><span class="NormalTextRun SCXW191537361 BCX0" data-ccp-parastyle="Normal (Web)"> using mostly the capabilities of built-in </span><span class="NormalTextRun SpellingErrorV2Themed SCXW191537361 BCX0" data-ccp-parastyle="Normal (Web)">zIIP</span><span class="NormalTextRun SCXW191537361 BCX0" data-ccp-parastyle="Normal (Web)"> engines, which cuts MSU costs, with </span><span class="NormalTextRun SCXW191537361 BCX0" data-ccp-parastyle="Normal (Web)">additional</span><span class="NormalTextRun SCXW191537361 BCX0" data-ccp-parastyle="Normal (Web)"> processing handled at a chosen cloud destination. In response to ETL&#8217;s high processing costs, many organizations have already moved the </span></span><span class="TextRun Underlined SCXW191537361 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW191537361 BCX0" data-ccp-charstyle="Hyperlink">data transformations</span></span><span class="TextRun SCXW191537361 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW191537361 BCX0" data-ccp-parastyle="Normal (Web)"> stage to the cloud to support analytics workloads and data lake creation.</span></span><span class="EOP Selected SCXW191537361 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h2><span class="TextRun SCXW236512981 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW236512981 BCX0" data-ccp-parastyle="heading 2">Why is ELT the </span><span class="NormalTextRun SCXW236512981 BCX0" data-ccp-parastyle="heading 2">better</span><span class="NormalTextRun SCXW236512981 BCX0" data-ccp-parastyle="heading 2"> path forward for mainframe organizations?</span></span></h2>
<p><span class="TextRun SCXW40199486 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW40199486 BCX0" data-ccp-parastyle="Normal (Web)">It would be wrong to oversimplify a decision between ETL and ELT—there are too many moving parts and decision points to weigh. But the key insight is this: ELT speaks to the evolving IT paradigms that ETL was never built to address.</span></span></p>
<p><span class="TextRun SCXW84825296 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW84825296 BCX0" data-ccp-parastyle="Normal (Web)">ELT is ideal for moving massive amounts of data to the cloud—typically to a data lake built to ingest any and all available data so that modern analytics can get to work.</span><span class="NormalTextRun SCXW84825296 BCX0" data-ccp-parastyle="Normal (Web)"> </span><span class="NormalTextRun SCXW84825296 BCX0" data-ccp-parastyle="Normal (Web)">That is why ELT is growing and making inroads specifically in the mainframe environment.</span><span class="NormalTextRun SCXW84825296 BCX0" data-ccp-parastyle="Normal (Web)"> ELT </span><span class="NormalTextRun SCXW84825296 BCX0" data-ccp-parastyle="Normal (Web)">represents</span><span class="NormalTextRun SCXW84825296 BCX0" data-ccp-parastyle="Normal (Web)"> </span><span class="NormalTextRun SCXW84825296 BCX0" data-ccp-parastyle="Normal (Web)">perhaps the</span><span class="NormalTextRun SCXW84825296 BCX0" data-ccp-parastyle="Normal (Web)"> best path to accelerating mainframe data movement to the cloud at scale, making ELT a key tool for IT organizations aiming at modernization </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW84825296 BCX0" data-ccp-parastyle="Normal (Web)">and at</span><span class="NormalTextRun SCXW84825296 BCX0" data-ccp-parastyle="Normal (Web)"> maximizing the value of their existing investments.</span></span></p>
<h2><span class="TextRun SCXW81460388 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW81460388 BCX0" data-ccp-parastyle="heading 2">Frequently asked questions</span></span></h2>
<p><strong>What is ETL and why has it been used in mainframe environments?</strong></p>
<p><span data-contrast="auto">ETL (extract, transform, and load) is a data integration method that pulls data from a source, transforms it to match a target schema, and loads it into a destination database. Mainframes have historically relied on ETL because it was well-suited to the structured, RDBMS-to-RDBMS data movement that dominated before cloud and big data workloads became standard.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What is the difference between ETL and ELT?</strong></p>
<p><span data-contrast="auto">ETL transforms data before loading it into the destination system. ELT loads raw data to the destination first and transforms it there, leveraging the destination system&#8217;s processing power. ELT is more flexible and better suited to modern cloud environments and data lakes than ETL.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>Why does ETL become a bottleneck for mainframe data migration?</strong></p>
<p><span data-contrast="auto">ETL requires all transformation to complete before data can move to its destination. At the scale of modern mainframe data volumes, this dependency on sequential transformation creates delays that make ETL impractical. ELT avoids the bottleneck by separating the load and transform steps entirely.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What is a zIIP engine and how does it reduce mainframe processing costs?</strong></p>
<p><span data-contrast="auto">A zIIP (IBM Z Integrated Information Processor) is a specialty engine on IBM mainframes designed to offload eligible workloads from general-purpose processors, reducing MSU charges. ELT workloads are often eligible for zIIP processing, making ELT significantly more cost-efficient than ETL for mainframe data migration projects.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>When does ETL still make sense over ELT?</strong></p>
<p><span data-contrast="auto">ETL remains appropriate for structured, database-to-database migrations where schema alignment and data quality transformation must happen before the data reaches its destination—particularly when data volumes are manageable and the target system requires pre-transformed data. For large-scale, cloud-bound mainframe modernization, ELT is generally the better choice.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><em><span class="TextRun SCXW249796767 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW249796767 BCX0" data-ccp-parastyle="Normal (Web)">The views and opinions expressed in this post are those of the author and do not necessarily reflect the official position of BMC.</span></span></em></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>EMA Names BMC a Value Leader in Workload Automation and Orchestration</title>
		<link>https://blogs.bmc.com/ema-radar-report-for-workload-automation/</link>
		
		<dc:creator><![CDATA[Basil Faruqui]]></dc:creator>
		<pubDate>Mon, 30 Mar 2026 13:00:44 +0000</pubDate>
				<category><![CDATA[Workload Automation Blog]]></category>
		<guid isPermaLink="false">http://www.bmc.com/blogs/?p=11782</guid>

					<description><![CDATA[<img width="700" height="400" src="https://s7280.pcdn.co/wp-content/uploads/2018/01/EMA-RADAR-Blog-Image-700x400px.jpeg.optimal.jpeg" class="attachment-large size-large wp-post-image" alt="" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2018/01/EMA-RADAR-Blog-Image-700x400px.jpeg.optimal.jpeg 700w, https://s7280.pcdn.co/wp-content/uploads/2018/01/EMA-RADAR-Blog-Image-700x400px-300x171.jpeg.optimal.jpeg 300w, https://s7280.pcdn.co/wp-content/uploads/2018/01/EMA-RADAR-Blog-Image-700x400px-24x14.jpeg.optimal.jpeg 24w, https://s7280.pcdn.co/wp-content/uploads/2018/01/EMA-RADAR-Blog-Image-700x400px-36x21.jpeg.optimal.jpeg 36w, https://s7280.pcdn.co/wp-content/uploads/2018/01/EMA-RADAR-Blog-Image-700x400px-48x27.jpeg.optimal.jpeg 48w" sizes="auto, (max-width: 700px) 100vw, 700px" />BMC has been named the overall highest performer and a Value Leader in the 2025 EMA Radar for Workload Automation and Orchestration—for the eighth consecutive time. Control-M also earned EMA’s recognition for Excellence in Mission-Critical Orchestration, reinforcing its standing as the market leader for enterprise-class orchestration. For organizations evaluating workload automation platforms, this recognition reflects […]]]></description>
										<content:encoded><![CDATA[<img width="700" height="400" src="https://s7280.pcdn.co/wp-content/uploads/2018/01/EMA-RADAR-Blog-Image-700x400px.jpeg.optimal.jpeg" class="attachment-large size-large wp-post-image" alt="" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2018/01/EMA-RADAR-Blog-Image-700x400px.jpeg.optimal.jpeg 700w, https://s7280.pcdn.co/wp-content/uploads/2018/01/EMA-RADAR-Blog-Image-700x400px-300x171.jpeg.optimal.jpeg 300w, https://s7280.pcdn.co/wp-content/uploads/2018/01/EMA-RADAR-Blog-Image-700x400px-24x14.jpeg.optimal.jpeg 24w, https://s7280.pcdn.co/wp-content/uploads/2018/01/EMA-RADAR-Blog-Image-700x400px-36x21.jpeg.optimal.jpeg 36w, https://s7280.pcdn.co/wp-content/uploads/2018/01/EMA-RADAR-Blog-Image-700x400px-48x27.jpeg.optimal.jpeg 48w" sizes="auto, (max-width: 700px) 100vw, 700px" /><p><span class="TextRun SCXW46953201 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW46953201 BCX0" data-ccp-parastyle="Normal (Web)">BMC has been named the overall highest performer and a Value Leader in the 2025 EMA Radar for Workload Automation and Orchestration—for the eighth consecutive time. Control-M also earned EMA&#8217;s recognition for Excellence in Mission-Critical Orchestration, reinforcing its standing as the market leader for enterprise-class orchestration. For organizations evaluating workload automation platforms, this recognition reflects Control-M&#8217;s maturity, breadth, and forward-looking platform strategy.</span></span></p>
<p><img loading="lazy" decoding="async" class="aligncenter size-full wp-image-15813" src="https://s7280.pcdn.co/wp-content/uploads/2025/10/EMA2025.png" alt="" width="1274" height="898" /></p>
<h2>Control-M: Leading Enterprise Orchestration</h2>
<p><span class="TextRun SCXW120213439 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW120213439 BCX0" data-ccp-parastyle="Normal (Web)">EMA&#8217;s report highlights Control-M&#8217;s unmatched maturity, scalability, and innovation cadence as the defining factors behind BMC&#8217;s Value Leader designation. With a modern, API-first </span><span class="NormalTextRun SCXW120213439 BCX0" data-ccp-parastyle="Normal (Web)">architecture and feature parity across SaaS and on-premises deployments, Control-M empowers organizations to orchestrate data pipelines, applications, and infrastructure with confidence and control.</span></span></p>
<h2><span lang="EN-US">Key Differentiators Called Out by EMA</span></h2>
<p><span class="NormalTextRun SCXW234311638 BCX0" data-ccp-parastyle="Normal (Web)">EMA </span><span class="NormalTextRun SCXW234311638 BCX0" data-ccp-parastyle="Normal (Web)">identified</span><span class="NormalTextRun SCXW234311638 BCX0" data-ccp-parastyle="Normal (Web)"> five areas where Control-M outperforms market averages across the EMA Radar workload automation evaluation criteria.</span></p>
<h3 aria-level="3"><span data-contrast="none">Mission-critical orchestration</span><span data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;335559738&quot;:160,&quot;335559739&quot;:80}"> </span></h3>
<p><span data-contrast="auto">Control-M leads in managing complex, infrastructure-intensive business processes with precision and reliability—the capability EMA specifically recognized with its Excellence in Mission-Critical Orchestration award.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3 aria-level="3"><span data-contrast="none">Hybrid and multi-cloud reach</span><span data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;335559738&quot;:160,&quot;335559739&quot;:80}"> </span></h3>
<p><span data-contrast="auto">Native integrations with AWS, Azure, GCP, Oracle Cloud, and Kubernetes enable seamless orchestration across hybrid and multi-cloud environments, without requiring separate tooling per platform.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3 aria-level="3"><span data-contrast="none">Data pipeline and DataOps leadership</span><span data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;335559738&quot;:160,&quot;335559739&quot;:80}"> </span></h3>
<p><span data-contrast="auto">Deep integrations with Snowflake, Databricks, Apache Airflow, and other modern data platforms make Control-M a cornerstone for enterprise data operations and DataOps workflows.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3 aria-level="3"><span data-contrast="none">DevOps and Jobs-as-Code</span><span data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;335559738&quot;:160,&quot;335559739&quot;:80}"> </span></h3>
<p><span data-contrast="auto">Developers can define, test, and promote workflows using JSON or Python, embedding workload orchestration directly into CI/CD pipelines and developer-native toolchains.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3 aria-level="3"><span data-contrast="none">Observability and AI</span><span data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;335559738&quot;:160,&quot;335559739&quot;:80}"> </span></h3>
<p><span data-contrast="auto">Embedded SLA clocks, anomaly detection, and Jett—Control-M&#8217;s fully integrated GenAI copilot—bring closed-loop intelligence to orchestration. Jett supports workflow optimization, SLA prediction, and real-time guidance, making AI a foundational part of the platform experience.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h2>A Vision for the Future: Orchestrator of Orchestrators</h2>
<p><span data-contrast="auto">EMA recognized BMC&#8217;s forward-looking strategy to position Control-M as the &#8220;orchestrator of orchestrators&#8221;—a unifying layer that spans ERP systems, DevOps pipelines, service management tools, and AI platforms. This vision is already taking shape through expanded integrations, enhanced Workflow Insights, and GenAI-powered Advisors.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">Control-M SaaS continues to gain momentum, offering global reach, hybrid visibility, and enterprise-grade resilience. With a single console view across SaaS and on-premises environments, organizations can modernize at their own pace without compromising governance or control.</span></p>
<h2><span class="TextRun SCXW151419870 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW151419870 BCX0" data-ccp-parastyle="heading 2">Strategic value drivers</span></span></h2>
<p><span class="TextRun SCXW189294394 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW189294394 BCX0" data-ccp-parastyle="Normal (Web)">BMC&#8217;s strategy in the workload automation and orchestration market rests on three core principles:</span></span></p>
<ul>
<li><span data-contrast="auto">End-to-end orchestration: Delivering orchestration of AI, data, and application workflows across hybrid environments—from multi-cloud to mainframe.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Agentic orchestration: Building on the GenAI-powered advisor Jett toward an agentic model, enabling a fleet of specialized AI agents to dynamically build, execute, and manage end-to-end workflows.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
<li><span data-contrast="auto">Flexible deployment: Providing SaaS or self-hosted options with a unified view that ensures consistency, governance, and control across environments.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></li>
</ul>
<p><span data-contrast="auto">These principles define how Control-M helps enterprises turn operational complexity into competitive advantage—operating with resilience and scaling innovation with confidence.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">To find out more about Control-M&#8217;s recognition and continued leadership in the EMA Radar workload automation space, </span><a href="/forms/ema-radar-report-workload-automation.html"><span data-contrast="none">download a copy of the report here</span></a><span data-contrast="auto">.</span></p>
<h2 aria-level="2"><span data-contrast="none">Frequently asked questions</span></h2>
<p><strong>What is the EMA Radar for Workload Automation and Orchestration?</strong></p>
<p><span data-contrast="auto">The EMA Radar for Workload Automation and Orchestration is an independent analyst evaluation by Enterprise Management Associates (EMA) that assesses vendors across criteria including functionality, architecture, deployment, cost advantage, and vendor strength. Value Leader status indicates the highest combined score for performance and value among evaluated vendors.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What does it mean that BMC was named a Value Leader in the EMA Radar?</strong></p>
<p><span data-contrast="auto">Being named a Value Leader means BMC&#8217;s Control-M achieved the highest overall performance score in EMA&#8217;s 2025 evaluation while also delivering strong cost-to-value positioning. BMC has held this designation for eight consecutive years, reflecting sustained leadership rather than a single-year result.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What is Excellence in Mission-Critical Orchestration?</strong></p>
<p><span data-contrast="auto">Excellence in Mission-Critical Orchestration is a specific EMA recognition awarded to vendors that demonstrate top-tier capability in managing complex, high-stakes, infrastructure-intensive workflows. Control-M earned this designation in the 2025 EMA Radar evaluation alongside its Value Leader status.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>How does Control-M support hybrid and multi-cloud workload automation?</strong></p>
<p><span data-contrast="auto">Control-M provides native integrations with major cloud platforms—AWS, Azure, GCP, and Oracle Cloud—as well as Kubernetes, enabling organizations to orchestrate workloads consistently across on-premises and cloud environments from a single unified console.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><em>The views and opinions expressed in this post are those of the author and do not necessarily reflect the official position of BMC.</em></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Operationalization and Orchestration: the Keys to Data Project Success</title>
		<link>https://blogs.bmc.com/operationalization-orchestration-keys-to-data-project-success/</link>
		
		<dc:creator><![CDATA[Basil Faruqui]]></dc:creator>
		<pubDate>Mon, 30 Mar 2026 10:18:31 +0000</pubDate>
				<category><![CDATA[Machine Learning & Big Data Blog]]></category>
		<guid isPermaLink="false">https://blogs.bmc.com/?p=52175</guid>

					<description><![CDATA[<img width="810" height="405" src="https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-1024x512.png" class="attachment-large size-large wp-post-image" alt="" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-1024x512.png 1024w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-300x150.png 300w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-768x384.png 768w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-810x405.png 810w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-1140x570.png 1140w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-24x12.png 24w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-36x18.png 36w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-48x24.png 48w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700.png 1400w" sizes="auto, (max-width: 810px) 100vw, 810px" />To operationalize data projects, organizations need automated, orchestrated, end-to-end visibility across every stage of the data pipeline. Without a structured operationalization framework, even the most sophisticated data initiatives stall before reaching production. A unified workflow orchestration platform that abstracts the complexity of disparate data tools is the critical bridge between a working prototype and enterprise-scale […]]]></description>
										<content:encoded><![CDATA[<img width="810" height="405" src="https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-1024x512.png" class="attachment-large size-large wp-post-image" alt="" decoding="async" loading="lazy" srcset="https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-1024x512.png 1024w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-300x150.png 300w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-768x384.png 768w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-810x405.png 810w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-1140x570.png 1140w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-24x12.png 24w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-36x18.png 36w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700-48x24.png 48w, https://s7280.pcdn.co/wp-content/uploads/2022/08/2-men-looking-at-screen-with-analytics_1400x700.png 1400w" sizes="auto, (max-width: 810px) 100vw, 810px" /><p><span class="TextRun SCXW183411053 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW183411053 BCX0" data-ccp-parastyle="Normal (Web)">To operationalize data projects, organizations need automated, orchestrated, end-to-end visibility across every stage of the data pipeline. Without a structured operationalization framework, even the most sophisticated data initiatives stall before reaching production. A unified workflow orchestration platform that abstracts the complexity of disparate data tools is the critical bridge between a working prototype and enterprise-scale deployment.</span></span></p>
<p><span class="NormalTextRun SCXW50851201 BCX0" data-ccp-parastyle="Normal (Web)">Data is abundant—and growing exponentially. But simply having the data </span><span class="NormalTextRun SCXW50851201 BCX0" data-ccp-parastyle="Normal (Web)">isn&#8217;t</span><span class="NormalTextRun SCXW50851201 BCX0" data-ccp-parastyle="Normal (Web)"> enough. Businesses consistently struggle to move data projects from development into production, and the cost of that gap is significant. In 2018, Gartner<sup>®</sup> predicted in their report entitled</span> “<a href="https://www.gartner.com/document/3894131?_ga=2.187073640.1560571058.1658333578-1905575251.1654693781">Predicts 2019: Artificial Intelligence Core Technologies</a>” <span class="NormalTextRun SCXW50851201 BCX0" data-ccp-parastyle="Normal (Web)">that only 15 percent of </span><span class="NormalTextRun SCXW50851201 BCX0" data-ccp-parastyle="Normal (Web)">cutting-edge</span><span class="NormalTextRun SCXW50851201 BCX0" data-ccp-parastyle="Normal (Web)"> data projects would make it into production by 2022—meaning 85 percent would fail to produce results. Separately, in it</span>s <a href="https://www.gartner.com/document/4012385?ref=solrAll&amp;refval=333171395">Top Trends in Data and Analytics, 2022 report</a>, <span class="NormalTextRun SCXW229335129 BCX0" data-ccp-parastyle="Normal (Web)">Gartner warned that organizations without </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW229335129 BCX0" data-ccp-parastyle="Normal (Web)">a sustainable</span><span class="NormalTextRun SCXW229335129 BCX0" data-ccp-parastyle="Normal (Web)"> data and analytics operationalization framework risk seeing their initiatives set back by up to two years.</span></p>
<h2><span class="TextRun SCXW120803163 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW120803163 BCX0" data-ccp-parastyle="heading 2">Why do data projects struggle to reach production?</span></span></h2>
<p><span class="TextRun SCXW206438030 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW206438030 BCX0" data-ccp-parastyle="Normal (Web)">The core challenge is scale. A project can work well in </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW206438030 BCX0" data-ccp-parastyle="Normal (Web)">prototype</span><span class="NormalTextRun SCXW206438030 BCX0" data-ccp-parastyle="Normal (Web)"> in one location, but if it </span><span class="NormalTextRun SCXW206438030 BCX0" data-ccp-parastyle="Normal (Web)">can&#8217;t</span><span class="NormalTextRun SCXW206438030 BCX0" data-ccp-parastyle="Normal (Web)"> be scaled nationally or globally, it has </span><span class="NormalTextRun SCXW206438030 BCX0" data-ccp-parastyle="Normal (Web)">essentially failed</span><span class="NormalTextRun SCXW206438030 BCX0" data-ccp-parastyle="Normal (Web)">.</span></span></p>
<p><span class="NormalTextRun SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)">As companies recognize the need to build operationalization into their plans, the industry has refocused on IT operations (</span><span class="NormalTextRun SpellingErrorV2Themed SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)">ITOps</span><span class="NormalTextRun SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)">)—generating a range of discipline-specific Ops frameworks: </span><span class="NormalTextRun SpellingErrorV2Themed SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)">DataOps</span><span class="NormalTextRun SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)"> for data, </span><span class="NormalTextRun SpellingErrorV2Themed SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)">MLOps</span><span class="NormalTextRun SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)"> for machine learning, AIOps for artificial intelligence, and </span><span class="NormalTextRun SpellingErrorV2Themed SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)">ModelOps</span><span class="NormalTextRun SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)"> for analytics modeling. This proliferation has even produced the catch-all term </span><span class="NormalTextRun SpellingErrorV2Themed SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)">XOps</span><span class="NormalTextRun SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)">—a placeholder, as some in the industry put it, for &#8220;we don&#8217;t know what&#8217;s coming </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)">next</span><span class="NormalTextRun SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)"> but it will involve Ops somehow.&#8221; The problem </span><span class="NormalTextRun SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)">isn&#8217;t</span><span class="NormalTextRun SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)"> ambition. </span><span class="NormalTextRun SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)">It&#8217;s</span><span class="NormalTextRun SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)"> </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)">complexity</span><span class="NormalTextRun SCXW202038360 BCX0" data-ccp-parastyle="Normal (Web)">.</span></p>
<h2><span class="TextRun SCXW234254430 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW234254430 BCX0" data-ccp-parastyle="heading 2">What are the four stages of a data pipeline?</span></span></h2>
<p><span class="TextRun SCXW174734770 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW174734770 BCX0" data-ccp-parastyle="Normal (Web)">Every data project shares the same four foundational stages, which form the building blocks of data pipelines:</span></span><span class="EOP Selected SCXW174734770 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3><span class="TextRun SCXW57269434 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW57269434 BCX0" data-ccp-parastyle="heading 3">1.Data ingestion</span></span></h3>
<p><span class="TextRun SCXW179630927 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW179630927 BCX0" data-ccp-parastyle="Normal (Web)">Orchestrating data from traditional sources such as enterprise resource planning (ERP) and customer relationship management (CRM) systems, financial platforms, and other systems of record—combined with data from devices, sensors, social media, weblogs, and IoT sensors and devices.</span></span></p>
<h3>2. <span class="TextRun SCXW124573789 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW124573789 BCX0" data-ccp-parastyle="heading 3">Data storage</span></span></h3>
<p><span class="TextRun SCXW75046186 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW75046186 BCX0" data-ccp-parastyle="Normal (Web)">Where and how data is stored depends significantly on persistence, the relative value of data sets, the rate of refresh for analytics models, and the speed at which data can move to processing.</span></span><span class="EOP Selected SCXW75046186 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3><span class="TextRun SCXW113036083 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW113036083 BCX0" data-ccp-parastyle="heading 3">3. Data processing</span></span></h3>
<p><span class="TextRun SCXW60862642 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW60862642 BCX0" data-ccp-parastyle="Normal (Web)">Processing requirements vary widely: How much </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW60862642 BCX0" data-ccp-parastyle="Normal (Web)">compute</span><span class="NormalTextRun SCXW60862642 BCX0" data-ccp-parastyle="Normal (Web)"> is needed? Is it constant or variable? Is workload scheduled, event-driven, or ad hoc? How are </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW60862642 BCX0" data-ccp-parastyle="Normal (Web)">costs</span><span class="NormalTextRun SCXW60862642 BCX0" data-ccp-parastyle="Normal (Web)"> minimized?</span></span><span class="EOP Selected SCXW60862642 BCX0" data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h3>4. <span class="TextRun SCXW122701519 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW122701519 BCX0" data-ccp-parastyle="heading 3">Insight delivery</span></span></h3>
<p><span class="TextRun SCXW52102706 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW52102706 BCX0" data-ccp-parastyle="Normal (Web)">The last mile—moving data output to analytics systems. The </span><span class="NormalTextRun SCXW52102706 BCX0" data-ccp-parastyle="Normal (Web)">insights</span><span class="NormalTextRun SCXW52102706 BCX0" data-ccp-parastyle="Normal (Web)"> layer is complex and shifts constantly as markets adopt </span><span class="NormalTextRun SCXW52102706 BCX0" data-ccp-parastyle="Normal (Web)">new technologies</span><span class="NormalTextRun SCXW52102706 BCX0" data-ccp-parastyle="Normal (Web)">. A new data analytics service that </span><span class="NormalTextRun SCXW52102706 BCX0" data-ccp-parastyle="Normal (Web)">isn&#8217;t</span><span class="NormalTextRun SCXW52102706 BCX0" data-ccp-parastyle="Normal (Web)"> in production at scale delivers no actionable insights and no business value, whether measured in revenue generation or operational efficiency.</span></span></p>
<h2><span class="TextRun SCXW16086127 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW16086127 BCX0" data-ccp-parastyle="heading 2">Why is end-to-end orchestration essential for data pipelines?</span></span></h2>
<p><span data-contrast="auto">The operational goal is to run data pipelines in a highly automated fashion with minimal human intervention and full visibility across every component. But almost every technology in the data pipeline comes with its own built-in automation utilities and tools—tools that are often not designed to work with each other. Stitching these together for end-to-end automation and orchestration is where teams hit the wall.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><span data-contrast="auto">This challenge has driven the rise of application and data workflow orchestration platforms: solutions that operate with speed and scale in production and abstract the underlying automation utilities so teams can focus on outcomes rather than infrastructure.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<h2><span class="TextRun SCXW17208362 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW17208362 BCX0" data-ccp-parastyle="heading 2">How does Control-M help operationalize data projects?</span></span><span class="EOP Selected SCXW17208362 BCX0" data-ccp-props="{&quot;134245418&quot;:true,&quot;134245529&quot;:true,&quot;335559738&quot;:160,&quot;335559739&quot;:80}"> </span></h2>
<div id="attachment_52176" style="width: 949px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" aria-describedby="caption-attachment-52176" class="wp-image-52176 size-full" src="https://s7280.pcdn.co/wp-content/uploads/2022/08/Gartner-Data-and-Analytics-Essentials.jpg.optimal.jpg" alt="" width="939" height="522" srcset="https://s7280.pcdn.co/wp-content/uploads/2022/08/Gartner-Data-and-Analytics-Essentials.jpg.optimal.jpg 939w, https://s7280.pcdn.co/wp-content/uploads/2022/08/Gartner-Data-and-Analytics-Essentials-300x167.jpg.optimal.jpg 300w, https://s7280.pcdn.co/wp-content/uploads/2022/08/Gartner-Data-and-Analytics-Essentials-768x427.jpg.optimal.jpg 768w, https://s7280.pcdn.co/wp-content/uploads/2022/08/Gartner-Data-and-Analytics-Essentials-810x450.jpg.optimal.jpg 810w, https://s7280.pcdn.co/wp-content/uploads/2022/08/Gartner-Data-and-Analytics-Essentials-24x13.jpg.optimal.jpg 24w, https://s7280.pcdn.co/wp-content/uploads/2022/08/Gartner-Data-and-Analytics-Essentials-36x20.jpg.optimal.jpg 36w, https://s7280.pcdn.co/wp-content/uploads/2022/08/Gartner-Data-and-Analytics-Essentials-48x27.jpg.optimal.jpg 48w" sizes="auto, (max-width: 939px) 100vw, 939px" /><p id="caption-attachment-52176" class="wp-caption-text">Figure 1. Gartner Data and Analytics Essentials: DataOps by Robert Thanaraj</p></div>
<p><span class="TextRun SCXW102543321 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW102543321 BCX0" data-ccp-parastyle="Normal (Web)">Control-M from BMC is an application and data workflow orchestration and automation platform that serves as the abstraction layer to simplify the complex data pipeline. Control-M enables end-to-end visibility and predictive service level agreements (SLAs) across any data technology or infrastructure, delivers data-driven insights in production at scale, and integrates </span><span class="NormalTextRun SCXW102543321 BCX0" data-ccp-parastyle="Normal (Web)">new technology</span><span class="NormalTextRun SCXW102543321 BCX0" data-ccp-parastyle="Normal (Web)"> innovations into even the most complex data pipelines.</span></span></p>
<p><span class="TextRun SCXW224159794 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW224159794 BCX0" data-ccp-parastyle="Normal (Web)">The Control-M platform offers a range of capabilities to automate and orchestrate application and data workflows:</span></span></p>
<ul>
<li>The <a href="/it-solutions/jobs-as-code.html">Control-M Automation API</a>, which promotes collaboration between Dev and Ops by allowing developers to embed production-ready workflow automation while applications are being developed.</li>
<li><a href="/it-solutions/control-m-integrations.html#cloud">Out-of-the-box support for cloud resources</a> including Amazon Web Services (AWS) Lambda and Azure Logic Apps, Functions, and Batch to help you leverage the flexibility and scalability of your cloud ecosystems.</li>
<li><a href="/it-solutions/control-m-managed-file-transfer.html">Integrated file transfers</a> with all your applications that allow you to move internal and external file transfers to a central interface to improve visibility and control.</li>
<li><a href="/it-solutions/control-m-capabilities.html#self-service">Self-Service</a> features that allows employees across the business to access the jobs data relevant to them.</li>
<li><a href="/documents/datasheets/control-m-application-integrator.html">Application Integrator</a>, which supports the creation of custom job types and deploys them in your Control-M environment quickly and easily.</li>
<li><a href="/it-solutions/control-m-capabilities.html#conversions">Conversion tools</a> that simplify conversion from third-party schedulers.</li>
</ul>
<p><span class="TextRun SCXW207872887 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW207872887 BCX0" data-ccp-parastyle="Normal (Web)">Data projects will continue to grow in strategic importance. Successfully operationalizing data workflows as a core part of project planning and execution is essential to business outcomes. </span><span class="NormalTextRun ContextualSpellingAndGrammarErrorV2Themed SCXW207872887 BCX0" data-ccp-parastyle="Normal (Web)">An</span><span class="NormalTextRun SCXW207872887 BCX0" data-ccp-parastyle="Normal (Web)"> </span></span><a class="Hyperlink SCXW207872887 BCX0" href="https://hub.bmc.com/what-is-application-workflow-orchestration" target="_blank" rel="noreferrer noopener"><span class="TextRun Underlined SCXW207872887 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW207872887 BCX0" data-ccp-charstyle="Hyperlink">application and data workflow orchestration</span></span></a><span class="TextRun SCXW207872887 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW207872887 BCX0" data-ccp-parastyle="Normal (Web)"> platform should be a foundational step in every </span><span class="NormalTextRun SpellingErrorV2Themed SCXW207872887 BCX0" data-ccp-parastyle="Normal (Web)">DataOps</span><span class="NormalTextRun SCXW207872887 BCX0" data-ccp-parastyle="Normal (Web)"> journey.</span></span></p>
<p>To learn more about how Control-M can help you find DataOps success, visit our <a href="/it-solutions/control-m.html">website</a>.</p>
<h2><span class="TextRun SCXW194426595 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="none"><span class="NormalTextRun SCXW194426595 BCX0" data-ccp-parastyle="heading 2">Frequently asked questions</span></span></h2>
<p><strong>What does it mean to operationalize a data project?</strong></p>
<p><span data-contrast="auto">To operationalize a data project means to move it from prototype or development into reliable, repeatable production at scale. Operationalization requires automating the core pipeline stages—ingestion, storage, processing, and insight delivery—so that data projects consistently produce actionable results without manual intervention.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>Why do most data projects fail to reach production?</strong></p>
<p><span data-contrast="auto">Most data projects fail to reach production because of complexity at scale. Each pipeline stage relies on different tools and technologies that don&#8217;t natively integrate, making end-to-end automation difficult to achieve. Without an orchestration layer to unify these components, teams cannot maintain the visibility and control needed to run pipelines reliably in production.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What is the difference between data automation and data orchestration?</strong></p>
<p><span data-contrast="auto">Data automation refers to executing individual tasks or processes without human intervention. Data orchestration coordinates multiple automated tasks across tools, systems, and environments to execute a complete data pipeline as a single managed workflow. Orchestration is what enables automation to work at enterprise scale.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What role does DataOps play in operationalizing data projects?</strong></p>
<p><span data-contrast="auto">DataOps applies DevOps principles—continuous integration, automation, and collaboration—to data pipeline development and operations. A DataOps framework helps organizations build operationalization into data projects from the start, reducing the time from prototype to production and improving the reliability of data-driven outcomes.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p><strong>What should I look for in a workflow orchestration platform?</strong></p>
<p><span data-contrast="auto">A workflow orchestration platform for data projects should provide end-to-end pipeline visibility, support for heterogeneous data technologies and cloud environments, event-driven and scheduled workload capabilities, integrated file transfer management, self-service access for business users, and APIs that enable Dev and Ops collaboration during development.</span><span data-ccp-props="{&quot;134233117&quot;:true,&quot;134233118&quot;:true,&quot;201341983&quot;:0,&quot;335559740&quot;:240}"> </span></p>
<p>GARTNER is a registered trademark and service mark of Gartner, Inc. and/or its affiliates in the U.S. and internationally and is used herein with permission. All rights reserved.</p>
<p><em><span class="TextRun SCXW80753407 BCX0" lang="EN-US" xml:lang="EN-US" data-contrast="auto"><span class="NormalTextRun SCXW80753407 BCX0" data-ccp-parastyle="Normal (Web)">The views and opinions expressed in this post are those of the author and do not necessarily reflect the official position of BMC.</span></span></em></p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
