• Skip to primary navigation
  • Skip to main content
Progressive Gauge

Progressive Gauge

Media and Research

  • Home
  • About
  • Blog
  • Projects, Samples and Items
  • Video
  • Contact Jack
  • Show Search
Hide Search

Data

The March of the Language Models

April 17, 2023 By Jack Vaughan

[April 17, 2023] – Had the opportunity to speak with Forrester Analyst Ronan Curran recently for a VentureBeat article. Of course, the topic was ChatGPT, generative AI, and Large Language Models.

His counsel was both optimistic and cautionary – a good summation of the bearings IT decision makers should set as they begin yet another tango with a new technology meme.

A handy summarizer-paraphraser tells me that Curran told VentureBeat that it would be a mistake to underestimate the technology, although it is still difficult to critically examine the many potential use cases for generative AI.

Yes, such applies to each technical challenge – every day. And it bears repeating as each new technology whispers or yells that the fundamental rules no longer apply – and yet they do.

Looking back on my conversation with Curran, I find insight in what some would say is obvious. The large language models are … large! And, as Curran told me, because they are large, they cost a lot to compute and train. This reminds us, as others have, that the LLM should be viewed like polo or horse racing – as a game for the rich.

Why do we say game for the rich? On one level, the LLM era stacks up as megacloud builders’ battle, albeit with aspects of the playground grudge match. Microsoft leader Satya Nadella, who had the thankless task of competing with Google on the search front, almost seems to chortle: “This new Bing will make Google come out and dance, and I want people to know that we made them dance.”

For the cloud giants, the business already had aspects of a war of attrition, as they staked data center regions across the globe. The folks at Semianalysis.com have taken a hard stab at estimating a day in the life of an LLM bean counter, and they suggest a “model indicating that ChatGPT costs $694,444 per day to operate in compute hardware costs.” Of course, these are back of the envelope estimates – and the titans that host LLMs will look to engineer savings.

The new LLM morning summons to mind a technology that  consumed  much attention not so long ago: Big Data. The magic of Hadoop had a difficult time jumping from the likes of Google, Facebook and Netflix to the broader market. Maybe Big Data should have been named ‘Prodigious Data’ – because that would have offered fairer warning to organizations that had to gather such data, administer it, and come up with clever and profitable use cases.

“What is Big Data good for?” was a common question, even in its heyday. Eventually the answer was “machine learning”.

Much of Big Data remained in the realm of the prototype. In the end, it was a step forward for enterprise analytics. Successes and failures alike came under the banner of prototyping. Clearly, experimentation is where we are now with ChatGPT.

The more interesting future for more people may lie in outcomes with small language models, Forrester’s Curran told me. These will succeed or fail on a use case by use case basis.

As industry observer Benedict Evans writes in “ChatGPT and the Imagenet moment,” ChatGPT feels like a step-change forward in the evolution of machine learning. It falls something short of sentience. There is potential but there are plenty of questions to answer before its arc can be well gauged.  [eof]

Read “Forrester: Question generative AI uses before experimentation” – VentureBeat Feb 24, 2023
https://venturebeat.com/ai/forrester-question-generative-ai-uses-before-experimentation/

Read “ChatGPT and the Imagenet moment” – ben-evans.com Dec 14, 2022
https://www.ben-evans.com/benedictevans/2022/12/14/ChatGPT-imagenet

The Inference Cost Of Search Disruption – Large Language Model Cost Analysis – Semianalysis.com Feb 9, 2023
https://www.semianalysis.com/p/the-inference-cost-of-search-disruption

Oracle gets the memo

October 24, 2022 By Jack Vaughan

Conf
Ellison updates Oracle Cloud World crowd.

THE SKEPTICAL ENQUIRER – Pride bordering on arrogation is a typical trait of successful enterprise tech companies – it’s definitely been the case with Oracle Corp. Many times it’s met the challenge of a changing computing paradigm with brio, taking its lead from irascible founder Larry Ellison, but the sum substance of its claims would be argued.

I may be a tea leaf reader here, but I see a subtle shift in Ellison’s irascibility in the minutes of last week’s Oracle Cloud World 2022 event.

Ellison’s lively quotes have long been a reporter’s friend. But it’s hard to relay his cloud computing pitches without adding footnotes. The consensus big cloud players are AWS, Google and Microsoft – while enterprise incumbents like Oracle, Teradata, and IBM are cited as tertiary at best.

Overall, Oracle’s financial growth has been modest in the last decade, while its cloud claims have been bold. It is by no means alone in the creative work it has done to define cloud, which is a promise of future growth, on its accounting ledger.

But, in terms of pitching everyday database automation advances of Oracle Autonomous Database [It’s enterprise blocking and tackling, right?] as the one-true self-driving database for the cloud of the future – well, Oracle has no equal there.

Its latest tactic is to boost its acquisition of health system giant Cerner as a great opportunity to rapidly modernize Cerner’s systems, move them to the Oracle Cloud, and count them as such on the financial report.

As I said, the company still experiences gradual overall gains, so I may be talking style rather than substance when I say they missed some big new tech boats, although they revisit these product line gaps from time to time. I’m talking about databases, development tools and persistence stratagems that all called for less spend and new thinking:

 

  • Oracle missed opportunities to expand its MySQL brand at a time when competitive PostgreSQL database versions were becoming go-to players in AWS and Microsoft stables for open-source distributed SQL. In September, the company came out with a bigger/better MySQL known as MySQL the Heatwave Database Service.
  • In fact, Oracle played down, ignored or glossed over inexpensive clustered open-source databases – both SQL and NoSQL — despite having the original Berkeley NoSQL stalwart in its hand.
  • JSON-oriented document database development led by MongoDB is a particular thorn. Oracle, like others, found ways to bring JSON to its RDBMS, but Mongo is still on the up stroke. Oracle last week addressed what it called a mismatch between JSON and the Oracle SQL database in the form of “JSON Relational Duality” – a new developer view supported in Oracle Database 23c.
  • Also, Oracle was slow to support cloud-friendly S3-style object storage. Not surprising in that the great goal is to place data into an Oracle database. But maybe it doesn’t have to be Oracle Autonomous Database. Last week, Oracle described MySQL Heatwave Lakehouse which may be a step in a broader direction of support.

This latter trait, object storage, seems to be getting a bit more of Oracle’s attention, as Snowflake Inc. rises on the back of its 3-in-1 distributed cloud data warehouse, data lake and data lake house. At Oracle’s yearly confab, Snowflake seemed to have garnered leader Ellison’s grudging admiration.

No small feat, that!

It’s a little hard to directly discern Ellison-speak-on-the-page. But his presentation moves salesfolks – and customers too. This recalls Curt Monash once calling him “one of the great public speakers and showmen of the era.”

What did Ellison tell the conference crowd? Well, besides a lot about how Oracle technology helped deliver Covid-19 vaccines, he spoke about the cloud market. People are moving from single-cloud to multi-cloud architectures, Ellison told the crowd at Oracle Cloud World 2022.

“The fact that this is happening is changing the behavior of technology providers…So the first thing that happened as people use multiple clouds is that service providers started deploying new services in multiple clouds, maybe most famously its Snowflake,” he said. Then, in a nod to the new highflier, he added, “And Oracle got the memo.”

“You know, we noticed, and we actually thought that was a good idea,” he said.

Of course, evolution to multi-cloud is a chance for Oracle to take another at bat in the cloud game. The bad news, as in the past, is that so much of the company’s effort is toward moving everything into the Oracle database. That is why any shifts to emphasize MySQL Heatwave would be notable.

Jack Vaughan is a writer and high-tech industry observer.

~~~~~~~~~~~~~~~~~~

My Back Pages
Data Lake Houses Join Data Lake in Big Data Analytics Race – IoT World Today – 2021
Purveyors of data lake houses and cloud data warehouses are many, and opinions on qualifications vary.

Oracle Cloud IaaS security gets superhero status from Ellison – TechTarget – 2018 [mp3 – Podcast]
It’s incremental!

Updated Oracle Gen 2 Cloud aims to challenge cloud leaders – TechTarget – 2018
The central product for Oracle is the database, and all the related tools that support the continued dependence of customers on the database.

On top of the RDB mountain – ADTmag 2001
Never lose the database sale: It’s tattooed on their foreheads.

How well can Nvidia tread the Agglomerverse?

September 25, 2022 By Jack Vaughan

Nvidia has worked hard to emerge from the worlds of graphic cards, gaming, and bitcoin mining to become a potent presence in enterprise AI considerations. It also is poised to play as a key vendor in the Metaverse, an AR-imbued but ill-defined repository for the next version of the Web.

More work is in store now as the GPU company – like most companies of any sort – navigates a more difficult economic environment – one where macro winds auger a possible enterprise spending slowdown. Already, Nvidia CEO Jensen Huang has led his crew into spaces others could not imagine.

Graphic Processing Units (GPUs) support ultrahigh memory bandwidth applications. They can churn through neural networks and sundry matrix multiplications like banshees. Huang and company pursue all their possible uses, and created a large portfolio of use cases, even as would-be competitors nip at their heels with more specialized offerings.

Visionary Huang, who we heard last week in keynotes and press conferences related to Nvidia’s GTC 2022 event, calls Nvidia an “Accelerated Computing Company.” And, he has set out to exploit “the Full Accelerated Computing Stack.”

These ambitions take form in a true slew of new offerings – ranging from the Nvidia DLSS3 deep learning sampler to GeForce RTX Series GPUs for neural rendering to Omniverse Cloud Services for Building and Operating Industrial Metaverse Applications, the Omniverse Replicator for synthetic data production and the 2,000-TFLOPS Thor SOC. The latter is probably well-described as “a super chip of epic proportions.”

Nvidia was early to see the possibility that AR/VR technology could drive a more interactive world-wide computing environment. The company coined it “the Omniverse” but now it’s joined others in the “metaverse” quest. For now, the metaverse is a loose agglomeration (the ‘Agglomerverse’?) of such elements as physics simulation, digital twins, and, of course, AI modeling. This puts Nvidia in competition or what Sam Alpert called coopetition with a host of other vendors. Hype vastly surpasses reality in today’s metaverse and the pay-off is both unclear and distant.

Meanwhile, Enterprise AI has found a place in data centers, and Nvidia has established a genuine foothold there. Obscured in the rush of GTC 2022 product announcements were less-than-flashy Apache Spark accelerator technology and AI inference announcements that may show up in revenue reports sooner than metaverse cases. Huang, for his part, sees the two technical domains playing off one another.

Be that as it may, in the metaverse and enterprise AI alike, Huang needs boots on the ground. These undertakings need great advances in skilling around big data.

It remains to be proved that corporations are anymore ready now to take on enterprise AI and the metaverse with imagination and execution. Can they imagine and execute on par or better than they did with Big Data Hadoop beginning ten years ago?

It’s worth noting that GTC 2022 software tools announcement were as proliferate as hardware news, showing the company is seeking ways to simplify the way to such advancements. Nvidia will likely need to take on greater headcount, and forge more mega-partnerships like one announced with Deloitte last week, if it going to successfully seed enterprise AI and metaverse apps.

Like most, Nvidia’s stock has been in free fall. But some of its challenges are unique. When US Government policy looked to slow down or block the transfer of advanced AI to China, Nvidia felt the brunt of it.

Meanwhile, the general rout of crypto currency impedes chip sales to crypto miners – and, as some news reports have it, a recent 2.0 update to the Ethereum blockchain takes a new proof-of-stake approach to processing and reduces the general call for GPUs for mining.

At the same time, the gaming card market has gone from famine to glut in the 24-month-plus period following the start of the global COVID pandemic. Moreover, the cost of these ever-bigger and more functional chips goes up-up-up, emptying gamer’s’ coffers.

Successes in these areas gave Nvidia wiggle room as it pursued enterprise AI. The wiggle room gets smaller just as the metaverse and enterprise AI to-do list gets taller. Among this week’s slew of portfolio additions there are some parts that will find users more quickly than others, and its up to Nvidia to suss those out and ensure they prosper. – Jack Vaughan

What’s it take to make #Metaverse real? [asks @deantak ]. In #GTC22 presser, Jensen discusses GDN – that is: a global Graphics Delivery Network – and notes as analog #Akamai Content Delivery Network (CDN). He said: “We have to put a new type of data center around the world.” pic.twitter.com/6Ur8IFwGJ3

— Jack Vaughan (@JackIVaughan) September 21, 2022

Jensen: We have a rich suite of domain specific application frame works. Now we need an army of experts to help customers apply these AI frameworks to automate their businesses. [Cue Deloitte soundtrack.] https://t.co/XBGewQGALP

— Jack Vaughan (@JackIVaughan) September 21, 2022

Omniverse Replicator — enables developers to generate physically accurate 3D synthetic data, and build custom synthetic-data generation tools to accelerate the training and accuracy of perception networks. https://t.co/t8HnVWvCcT

— Jack Vaughan (@JackIVaughan) September 20, 2022

InfluxDB Time-Series Cloud Launches on Google Cloud

February 13, 2020 By Jack Vaughan

We spoke recently with open-source time-series database maker InfluxData Inc. The occasion was the formal launch of InfluxDB Cloud — a database as a service — on Google Cloud. AWS support precedes this, and Microsoft Azure support is next up.
Launches like this are closely watched as the cloud database has become a proving ground for the future of the open source database. Highly visible wrangling between MongoDB and AWS, particularly, have placed the issue in bas relief.
The questions arise:

Will powerful cloud providers exploit open source loop holes to co-opt small startups database innovations?

Are the innovators, in the words of the old Sonny Boy Williamson song, simply “fattening frogs for snakes?”

Let’s put blues philosophy aside for the moment. Let’s look at the record.

For its part, Google has put increasing effort into closing the wide gap between itself and cloud leaders AWS and Microsoft. This mercantile motive has led to altruism that sees Google taking an interest in effectively partnering with open-source database wunderkinds.
This was clear last year as Google seized on AWS’s wranglings (with Elastic, MongoDB, principally) when it very visibly announced strategic partnerships with select open source database makers including InfluxData, MongoDB, Neo4j, Redis Labs and others. That is some of the background behind the February 4 InfluxDB Cloud on Google Cloud roll out.

For InfluxData CEO Evan Kaplan there is no question that the future of the database, including the InfluxDB time-series database, is in the cloud.
“The market has voted on open source databases,” he said. “We all recognize that the next generation of database applications are going to be founded on open source, and they will be offered on the cloud.”
Clearly, Kaplan and other opensource database company leaders walk the delicate but familiar line of “co-opitition” – working with big cloud providers to reach an audience, while also working to best the big players in ease-of-use, features and performance.
“We have to be able to compete with our technology on the cloud,” he told us.

Databases on the cloud are now somewhere in a multiyear evolution that is far from complete. But, with more and more data accruing to the cloud and the complexity of managing massive infrastructure growing, it makes sense when Kaplan says that the future of databases is in the cloud.
That could be especially true for time series data and analytics in the Industrial Internet of Things (IIoT). That is one of the sweet spots for InfluxData’s technology, Kaplan adjudges. Time series analytics have special currency in IIoT because the trends of blips of data over time disclose important information on how systems and their environments are changing. Such is part and parcel of the Industrial Internet of Things (IIoT).
While IIoT is a perhaps fancy term, it does stand for something. That is, the next generation of industrial systems that adjust their activity in real-time based on immediate analysis of data points.
Today industry looks at data through a new lens, with the purpose of improving operations. Some of these industries have been doing time series data analysis for quite a while. The speed of processing, the amount of data, the ease of programming are all elements that make this a promising technology area, and these are features vendors will focus on going forward.
As with the move from the mainframe to client-server architectures in the 1990s, this could be a telling moment for industrial systems. That is because the increasingly massive scale of processing on the cloud is accompanied by an architectural shift to serverless architecture.
Whether you select a big cloud provider or a smaller database player, the new architecture sees the advent of serverlessness. Database systems run as instances for customers – they run as instances on demand, like utilities (which, after all, was one of the first bits of nomenclature that rose up to describe what later became known as ‘the cloud’).
“InfluxDB Cloud is not tied to a server,” explains Kaplan. “It’s a serverless time-series platform built to collect, store and query processes, and to visualize rapidly ingested raw, high-precision time-stamped data.” The performance thereto is the secret sauce InfluxDB hopes will keep it a few steps ahead of the influential cloud players that dominate areas of computing today.

Other time-series specialists such as Graphite, TimeScale and others are in that hunt as well, while database stalwarts Oracle and IBM continue to tune their top offerings to bring time-series tech to a broader group of their customers. – Jack Vaughan

Progressive Gauge

Copyright © 2023 · Jack Vaughan · Log in

  • Home
  • About
  • Blog
  • Projects, Samples and Items
  • Video
  • Contact Jack