Weblogic Authors: Yeshim Deniz, Elizabeth White, Michael Meiner, Michael Bushong, Avi Rosenthal

Related Topics: Weblogic

Weblogic: Article

Acid Reign

The transaction processing monitor is dead; long live the transaction processing monitor

In browsing around the Web, as one occasionally does in a free nanosecond, I read an interesting article about twp-phase commit transactions by Gregor Hohpe of ThoughtWorks ("Your Coffee Shop Does Not Use Two Phase Commit"). Gregor comes at the subject from the direction opposite the one I usually take in this column, since I am of a TP persuasion, but he covers the same arguments that I have explored in the past and comes to similar conclusions.

"Your Coffee Shop Doesn't Use Two-Phase Commit": Should You?
Briefly, two-phase commit can be costly from a performance standpoint - in terms of both end-to-end transaction time and throughput - and you need to make sure you cost-justify your decision to use it when you do. On a finer point of detail, I would take issue with the implication in the article that there is a straight choice between synchronous processing and two-phase commit - most of the transaction systems I have ever seen involve access to one database and one asynchronous reliable queue in the context of a single transaction - thus guaranteeing that some updates happen and a message is guaranteed to be delivered to the next step in the process at some time in the future as an atomic unit, but that's another story.

So, back from my digression, the reason I wanted to write a follow-on to this article was because it seemed to me that the article illustrates a good point about the benefits of transactions and transaction management - particularly in the broadest sense of the term. To briefly summarize it so that the rest of my article makes sense, the thesis is that Starbucks uses an asynchronous model for accepting coffee requests into its main business system (the barista) in order to maximize the potential throughput of coffee from the shop, and hence maximize revenue. The potential cost of this optimization to the "happy day" scenario - the assumption that all is going forward with no errors - is the odd need to pause and hold up the line while wrong drinks are disposed of and remade, or money is refunded to unhappy (and still thirsty) punters. Indeed, if this is the only cost of this assumption, then the asynchronous case is clearly the correct design - throughput is maximized and the cost of unwinding the odd failure is outweighed by the less costly, less complex system we have built.

However, back in the real world away from analogy land, this is often a better "project phase one" argument than a "deployment lifetime" argument. To return to the analogy by way of illustration, imagine that our coffee production line is now in place and we are happily raking in money for strangely named coffees, and suddenly we get a good idea: How about improved quality of service for regular customers? For those really important coffee drinkers - rendered hyper-impatient by the caffeine they are wired with - we want to tell them how long it will be until they can expect their coffee. This poses a problem for us. From the moment a name got written on the empty coffee cup and it got queued on top of the coffee machine, we lost track of it. We are relying on the customer to hang about and wait until his name is called before customer and cup (now replete with coffee) are reunited. To track the cup in the queue, we will need to adopt a new strategy - maybe we have a yellow cup for the priority customer, so we can see the yellow cups processing up the queue. Well, that's fine until we have too many privileged customers, when we will no longer be able to distinguish which yellow cup we are looking for. What now? Suddenly we start to wish that we had a nice synchronous coffee production process, so we knew where we stood; of course, the analogy is starting to creak here a bit - the time it takes to produce a coffee is relatively short, and there is only one business system in the process, so this is not really that great an issue (which is a pity, because I was just about to propose attaching RFID tags to each cup, to allow us to associate them with their intended recipients…), but there is a core of truth here.

Attach an RFID Tag to Each Cup
Because of the apparent simplicity of asynchronous queuing-based systems, they are widely deployed. Additionally, a widely felt pain they create is the lack of visibility into "in-flight" business transactions. There is a clear and present demand for "executive dashboard" type facilities, so that managers have some insight into how the systems for which they are responsible are running - and some ability to foresee and forestall problems. This usually causes the nice simple phase one MOM-based systems to be glued to a lot of information gathering infrastructure (usually, more queues) with some kind of event correlation machinery on the back end to give an indication of what is passing through the system, and what looks unusual or possibly erroneous. Where did that original elegant simplicity go?

One oft-overlooked advantage of building systems with a transaction manager coordinating them is that we get an out-of-the-box central place that we can go to see which business event has touched which resource, and what the outcome is expected to be.

Of course, that said we still do have the bottleneck that the enforcement of ACID properties places on our throughput (chiefly, because of the database locks that ACID implies). It is here that I can mount another of my favorite hobbyhorses. One way to get the benefits of central coordination of transactions (in the loose sense, simply meaning correlated business activities) without incurring the penalties of data locking and contention is to relax the (technical) strictures of the ACID rules and allow a more easygoing, business reality-focused view of how the transactions correlate the business events. This is the idea behind a "next-generation" transaction concept such as cohesions, which first surfaced in the OASIS BTP standard and are now informing the debate within the WSBusinessActivity proposed standard.

Have Your Coffee and Drink It Too
Maybe, when these transaction standards mature, we will at last be able to have our coffee and drink it too?


More Stories By Peter Holditch

Peter Holditch is a senior presales engineer in the UK for Azul Systems. Prior to joining Azul he spent nine years at BEA systems, going from being one of their first Professional Services consultants in Europe and finishing up as a principal presales engineer. He has an R&D background (originally having worked on BEA's Tuxedo product) and his technical interests are in high-throughput transaction systems. "Of the pitch" Peter likes to brew beer, build furniture, and undertake other ludicrously ambitious projects - but (generally) not all at the same time!

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.

IoT & Smart Cities Stories
Business professionals no longer wonder if they'll migrate to the cloud; it's now a matter of when. The cloud environment has proved to be a major force in transitioning to an agile business model that enables quick decisions and fast implementation that solidify customer relationships. And when the cloud is combined with the power of cognitive computing, it drives innovation and transformation that achieves astounding competitive advantage.
Machine learning has taken residence at our cities' cores and now we can finally have "smart cities." Cities are a collection of buildings made to provide the structure and safety necessary for people to function, create and survive. Buildings are a pool of ever-changing performance data from large automated systems such as heating and cooling to the people that live and work within them. Through machine learning, buildings can optimize performance, reduce costs, and improve occupant comfort by ...
René Bostic is the Technical VP of the IBM Cloud Unit in North America. Enjoying her career with IBM during the modern millennial technological era, she is an expert in cloud computing, DevOps and emerging cloud technologies such as Blockchain. Her strengths and core competencies include a proven record of accomplishments in consensus building at all levels to assess, plan, and implement enterprise and cloud computing solutions. René is a member of the Society of Women Engineers (SWE) and a m...
Early Bird Registration Discount Expires on August 31, 2018 Conference Registration Link ▸ HERE. Pick from all 200 sessions in all 10 tracks, plus 22 Keynotes & General Sessions! Lunch is served two days. EXPIRES AUGUST 31, 2018. Ticket prices: ($1,295-Aug 31) ($1,495-Oct 31) ($1,995-Nov 12) ($2,500-Walk-in)
According to Forrester Research, every business will become either a digital predator or digital prey by 2020. To avoid demise, organizations must rapidly create new sources of value in their end-to-end customer experiences. True digital predators also must break down information and process silos and extend digital transformation initiatives to empower employees with the digital resources needed to win, serve, and retain customers.
IoT is rapidly becoming mainstream as more and more investments are made into the platforms and technology. As this movement continues to expand and gain momentum it creates a massive wall of noise that can be difficult to sift through. Unfortunately, this inevitably makes IoT less approachable for people to get started with and can hamper efforts to integrate this key technology into your own portfolio. There are so many connected products already in place today with many hundreds more on the h...
Digital Transformation: Preparing Cloud & IoT Security for the Age of Artificial Intelligence. As automation and artificial intelligence (AI) power solution development and delivery, many businesses need to build backend cloud capabilities. Well-poised organizations, marketing smart devices with AI and BlockChain capabilities prepare to refine compliance and regulatory capabilities in 2018. Volumes of health, financial, technical and privacy data, along with tightening compliance requirements by...
Charles Araujo is an industry analyst, internationally recognized authority on the Digital Enterprise and author of The Quantum Age of IT: Why Everything You Know About IT is About to Change. As Principal Analyst with Intellyx, he writes, speaks and advises organizations on how to navigate through this time of disruption. He is also the founder of The Institute for Digital Transformation and a sought after keynote speaker. He has been a regular contributor to both InformationWeek and CIO Insight...
Digital Transformation is much more than a buzzword. The radical shift to digital mechanisms for almost every process is evident across all industries and verticals. This is often especially true in financial services, where the legacy environment is many times unable to keep up with the rapidly shifting demands of the consumer. The constant pressure to provide complete, omnichannel delivery of customer-facing solutions to meet both regulatory and customer demands is putting enormous pressure on...
Andrew Keys is Co-Founder of ConsenSys Enterprise. He comes to ConsenSys Enterprise with capital markets, technology and entrepreneurial experience. Previously, he worked for UBS investment bank in equities analysis. Later, he was responsible for the creation and distribution of life settlement products to hedge funds and investment banks. After, he co-founded a revenue cycle management company where he learned about Bitcoin and eventually Ethereal. Andrew's role at ConsenSys Enterprise is a mul...