The Cloud-Ready Mainframe: Extending Your Data’s Reach and Impact

(This post is sponsored by VirtualZ Computing)

Some of the largest enterprises are finding new uses for their mainframes. And instead of competing with cloud and distributed computing, the mainframe has become a complementary asset that adds new productivity and a level of cost-effective scale to existing data and applications. 

While the cloud does quite well at elastically scaling up resources as application and data demands increase, the mainframe is purpose-built for the largest-scale digital applications. But more importantly, it has kept pace as these demands have mushroomed over its 60-year reign, and why so many large enterprises continue to use them. Having them as part of a distributed enterprise application portfolio could be a significant and savvy use case, and be a reason for increasing their future role and importance.

Estimates suggest that there are about 10,000 mainframes in use today, which may not seem a lot except that they can be found across the board in more than two-thirds of Fortune 500 companies, In the past, they used proprietary protocols such as Systems Network Architecture, had applications written in now-obsolete coding languages such as COBOL, and ran on custom CPU hardware. Those days are behind us: instead, the latest mainframes run Linux and TCP/IP across hundreds of multi-core  microprocessors. 

But even speaking cloud-friendly Linux and TCP/IP doesn’t remove two main problems for mainframe-based data. First off, many mainframe COBOL apps are their own island, isolated from the end-user Java experience and coding pipelines and programming tools. To break this isolation usually means an expensive effort to convert and audit the code. 

A second issue has to do with data lakes and data warehouses. These applications have become popular ways that businesses can spot trends quickly and adjust IT solutions as their customer’s data needs evolve. But the underlying applications typically require having near real-time access to existing mainframe data, such as financial transactions, sales and inventory levels or airline reservations. At the core of any lake or warehouse is conducting a series of “extract, transform and load” operations that move data back and forth between the mainframe and the cloud. These efforts only transform data at a particular moment in time, and also require custom programming efforts to accomplish.

What was needed was an additional step to make mainframes easier for IT managers to integrate with other cloud and distributed computing resources, and that means a new set of software tools. The first step was thanks to initiatives such as the use of IBM’s z/OS Connect. This enabled distributed applications to access mainframe data. But it continued the mindset of a custom programming effort and didn’t really provide direct access to distributed applications.

To fully realize the vision of mainframe data as equal cloud nodes required a major makeover, thanks to companies such as VirtualZ Computing. They latched on to the OpenAPI effort, which was previously part of the cloud and distributed world. Using this protocol, they created connectors that made it easier for vendors to access real-time data and integrate with a variety of distributed data products, such as MuleSoft, Tableau, TIBCO, Dell Boomi, Microsoft Power BI, Snowflake and Salesforce. Instead of complex, single-use data transformations, VirtualZ enables real-time read and write access to business applications. This means the mainframe can now become a full-fledged and efficient cloud computer. 

VirtualZ CEO Jeanne Glass says, “Because data stays securely and safely on the mainframe, it is a single source of truth for the customer and still leverages existing mainframe security protocols.” There isn’t any need to convert COBOL code, and no need to do any cumbersome data transformations and extractions.

The net effect is an overall cost reduction since an enterprise isn’t paying for expensive high-resource cloud instances. It makes the business operation more agile, since data is still located in one place and is available at the moment it is needed for a particular application. These uses extend the effective life of a mainframe without having to go through any costly data or process conversions, and do so while reducing risk and complexity. These uses also help solve complex data access and report migration challenges efficiently and at scale, which is key for organizations transitioning to hybrid cloud architectures. And the ultimate result is that one of these hybrid architectures includes the mainframe itself.

Building the world’s largest digital camera

LSST Camera and SLAC Camera Team-5.jpgThe world’s largest digital camera is a 3200 megapixel behemoth that sits on top of a mountain observatory complex in Chile. Ironically, it was created by engineers that in the past have focused on tracking the universe’s smallest subatomic particles. The camera has one acronym (LSST) and goes by two long names — the legacy survey of space and time, or Large-Aperture Synoptic Survey Telescope.
The camera is part of a telescope at the Vera Rubin observatory, named after an American astronomer that studied dark matter. Everything is still under construction and is expected to become operational next year. When it does, it will work in very different ways than its peers. And while the Webb telescope has gotten plenty of press for flying around the sun these past couple of years — and rightly so, I don’t want to diminish its accomplishments — the  Simonyi telescope at the Rubin is an interesting science tool in its own right. And yes, that name is familiar to many of you. Charles Simonyi worked in the early years at Microsoft, and both he and Bill Gates were early donors to the project.
The Rubin project has been long in coming, just like the Webb. In fact, pieces of it were built in different factories and labs around the world. The camera came from California (the Stanford Linear Accelerator team), the mount was from Spain, and Chile put together the buildings housing everything.
First off, if you have in your mind this is a place where astronomers go to peer through the eyepiece of the telescope and stare at the night sky, put that picture aside. This is a digital camera, and it operates hands-off for the most part. Its goal is nothing short of extraordinary. Every three nights, weather permitting, it photographs the entire night sky, moving around in a pre-programmed pattern. Most telescopes of the past were firmly anchored to their mountain top aeries.
In the past, telescopes like Webb and other expensive instruments required scientists to schedule time on them to focus on particular areas of the sky, and then download what was collected. Committees would vet proposals and schedule the sessions accordingly. Having a telescope that sweeps the entire sky — and doing it in such high resolution — means that you can approach observations in a completely different way.
First off, you don’t download anything. Given the size of the datasets, that would take time, even on high speed bandwidth. All the data stays intact, and you run your queries remotely.
This is a massive amount of data — petabytes worth — and it is all uploaded to an open source repository. Anyone can access the information for their research or just for curiosity. I imagine that schools will jump on board using this archive. It might change the way we teach astronomy and it certainly will reach a wider audience.
Also, the science team behind the Rubin is developing software that mimics what the early astronomers did manually, to seek out changes in the observations. Did a planet move in front of its star? Is a black hole forming someplace? I remember as a child reading about Clyde Tombaugh and how he discovered Pluto (poor Pluto, now downgraded to a demi-planet) in 1930 by looking at photographs taken on different nights to find its movement. He used a device called a blink microscope to quickly flip back and forth between the two photographic plates. Now we have open source code to do that tedious task.
This means that discoveries will be made almost every night, because the universe is a busy place. Scientists don’t have to depend on picking the right time and piece of sky real estate to observe a supernova, but can have software seek out the possible event.
Another distinction: unlike the infrared-based Webb, Rubin operates in visible light.
Finally, what I also liked is that the project is the first time a publicly-funded astronomy effort has been named for a woman.

How to make AI models more processor efficient

I was amused to read that a mathematical method that I first learned as an undergraduate has been found to help make AI models more processing efficient. The jump is pretty significant, if the theories hold in practice: a drop in 50x power consumption. This translates into huge cost savings: some estimate that the daily electric bill for running ChatGPT v3.5 is $700,000.

The method is called matrix multiplication and you can find a nice explanation here if you really want to learn what it is. MM is at the core of many mathematical models, and while I was in school we didn’t have the kind of computers (or built-in to our digital spreadsheets or in Python code) to make this easier, so we had to do these by hand as we were walking miles uphill to and from school in the snow.

MM dates back to the early 1800s when a French mathematician Jacques Binet figured it out. It became the foundational concept of linear algebra, something taught to math, engineering and science majors early on in their college careers.

The researchers figured out that, with the right custom silicon, they could run a billion-parameter model for about 13 watts. How do you make the connection between the AI models and MM? Well, your models are using words, and each word is represented by some random number, which are then organized into matrices. You do the MM to create phrases and figure out the relationships between adjacent words. Sounds easy, no?

Well, imagine that you have to do these multiplications a gazillion times. That adds up to a lot of processing. The researchers figured out a clever way to reduce the multiplications to simple addition, and then designed a special chipset that was optimized accordingly for these operations.

It is a pretty amazing story, and just shows you the gains that AI is making literally at the speed of light. It also shows you how some foundational math concepts are still valid in the modern era.

SiliconANGLE: Biden’s AI executive order is promising, but it may be tough for the US to govern AI effectively

President Biden signed a sweeping executive order yesterday covering numerous generative AI issues, and it’s comprehensive and thoughtful, as well as lengthy.

The EO contains eight goals along with specifics of how to implement them, which on the surface sounds good. However, it may turn out to be more inspirational than effective, and it has a series of intrinsic challenges that could be insurmountable to satisfy. Here are six of my top concerns in a post that I wrote for SiliconANGLE today.

All in all, the EO is still a good initial step toward understanding AI’s complexities and how the feds will find a niche that balances all these various — and sometimes seemingly contradictory — issues. If it can evolve as quickly as generative AI has done in the past year, it may succeed. If not, it will be a wasted opportunity to provide leadership and move the industry forward.

Using Fortnite for actual warfare

What do B-52s and a Chinese soccer stadium have in common? Both are using Epic Games’ Unreal Engine to create digital twins to help with their designs. Now, you might think having a software gaming engine would be a stretch to retrofit the real engines on a 60-plus year old bomber, but that is exactly what Boeing is doing. The 3D visualization environment makes it easier to design and provide faster feedback to meet the next generation of military pilots.

This being the military, the notion of “faster” is a matter of degree. The goal is for Boeing to replace the eight Pratt and Whitney engines on each of 60-some planes, as well as update cockpit controls, displays and other avionics. And the target date? Sometime in 2037. So check back with me then.

Speaking of schedules, let’s look at what is happening with that Xi’an stadium. I wrote about the soccer stadium back in July 2022 and how the architects were able to create a digital twin of the stadium to visualize seating sight lines and how various building elements would be constructed. It is still under construction, but you can see a fantastic building taking shape in this video. However slowly the thing is being built, it will probably be finished before 2037, or even before 2027.

Usually, when we talk about building digital twins, we mean taking a company’s data and making it accessible to all sorts of analytical tools. Think of companies like Snowflake, for example, and what they do. But the gaming engines offer another way to duplicate all the various systems digitally, and then test different configurations by literally putting a real bomber pilot in a virtual cockpit to see if the controls are in the right place, or the new fancy hardware and software systems can provide the right information to a pilot. If you look at the cockpit of another Boeing plane — the iconic 747, now mostly retired, you see a lot of analog gauges and physical levers and switches.

Now look at the 777 cockpit — see the difference? Everything is on a screen.

product image

It is ironic in a way: we are using video gaming software to reproduce the real world by placing more screens in front of the people that are depicted in the games. A true Ender’s Game scenario, if you will.

SiliconANGLE: Smarter shopping carts are coming but usability and privacy concerns loom

A new version of the smarter shopping cart will be coming to a nearby market this fall. Thanks to various partnerships and technological innovations of Instacart Inc., the latest embodiment of what the company calls Caper Carts will be able to track purchases while shoppers navigate through the aisles. The goal is to make it easier for shoppers to skip the checkout lines.

But it’s a tough reach, given the complexities of the retail channel and how the items will be scanned and tracked. If it works, it could be a major time saver. If it stumbles, it could be another example of bad user interface technology that is presently in most grocery and other retail chains: automated checkout scanning lanes. I write about it for SiliconANGLE today here.

 

SiliconANGLE: It won’t be long before we are all chatbot prompt engineers

Back in January, Andrej Karpathy, who now works for OpenAI LP and used to be the director of artificial intelligence for Tesla Inc., tweeted: “The hottest new programming language is English.” Karpathy was only semiserious, yet he has identified a new career path: AI chatbot prompt engineer. It could catch on.

The term describes the people who create and refine the text prompts that users type into the chatbot query windows — hence the use of English, or any other standard human language. These types of engineers don’t need to learn any code, but they do need to learn how the AI chatbots work, what they’re good at doing and what they’re not good at doing.

I interviewed several experts about whether the discipline will become its own career path in my post for SiliconANGLE here.

SiliconANGLE: ChatGPT detectors still have trouble separating human and AI-generated texts

The growth of ChatGPT and other chatbots over the past year has also stimulated the growth of software that can be used to detect whether a text is most likely to originate from these automated tools. That market continues to evolve, but lately there is some mixed news that not all detector programs are accurate, and at least one has actually been discontinued.

I examine two different academic reviews of several of these detector tools, and how they have failed under varying circumstances, for my post for SiliconANGLE here.

SiliconANGLE: That next computer in the cloud could be an IBM mainframe

A small Minneapolis mainframe computer software startup is poised to change the way enterprises use and share data across the cloud.

Virtual Z Computing Inc. claims to be the first and only women-founded and women-led mainframe systems integrator in history. That is a bold position, but perhaps more important is its pair of revolutionary software applications called Lozen and Zaac that connect native mainframe data with various third-party distributed, cloud-based applications.

I explain how the company’s products fit into the future of cloud computing in this story for SiliconANGLE here. 

SiliconANGLE: Databases then and now: the rise of the digital twin

When I first started in IT, back in the Mainframe Dark Ages, we had hulking big databases that ran on IBM’s Customer Information Control System, written in COBOL. These mainframes ran on a complex collection of hardware and operating systems that was owned lock, stock, and bus and tag barrel by IBM. The average age of the code was measured in decades, and code changes were measured in months. They contained millions of transactions, and the data was always out of date since it was a batch system, meaning every night new data would be uploaded.

Contrast that to today’s typical database setup. Data is current to the second, code is changed hourly, and the nature of what constitutes a transaction has changed significantly to something that is now called a “digital twin,” which I explain in my latest post for SiliconANGLE here.

Code is written in dozens of higher-level languages that have odd names that you may never have heard of, and this code runs on a combination of cloud and on-premises equipment that uses loads of microprocessors and open source products that can be purchased from hundreds of suppliers.

It really is remarkable, and that these changes have happened all within the span of a little more than 35 years. You can read more in my post.