• Skip to primary navigation
  • Skip to main content
Progressive Gauge

Progressive Gauge

Media and Research

  • Home
  • About
  • Blog
  • Projects, Samples and Items
  • Video
  • Contact Jack
  • Show Search
Hide Search

Archives for May 2024

Cadence discusses AI-driven fit-for-purpose chips

May 22, 2024 By Jack Vaughan

Phonautograph

The era of hyperscalers designing their own fit-for-purpose chips began with a series of AI chips from Google, Microsoft and AWS. Others cite the work of Apple and others to forge their own destinies in custom-chip designs for smartphones.

The trend has continued, but it is not clear when or if it will spread to other ranks among Information Technology vendors.

The chips specifically built to run the big players’ finely honed AI algorithms are, for now, the sweet spot for fit-for-purpose.

The surge in interest in in-house chip designs got rolling a few years ago with Google’s Tensor Processing Unit, which is specially tuned to meet Google’s AI architecture. The search giant has followed that with the Argos chip, meant to speed YouTube video transcoding, and Axion, said to drive data center energy efficiencies.

Chip design for Google and its ilk is enabled by deep pockets of money. The big players have ready mass markets that can justify big expenses in IC design staff and resources as well.

Chief among those resources is Electronic Design Automation tooling from the likes of Cadence Systems.  This week, Anirudh Devgan, President & CEO of Cadence, discussed the trend at the J.P. Morgan 52nd Global Technology, Media and Communications Conference in Boston.

He said the key reasons companies go the self-designed route are: Achieving domain-specific product differentiation, gaining control over the supply chain and production schedule, and realizing cost benefits at scale when their chip shows it will find use at sufficient volume.

Domain-specific differentiation allows companies to a create chip tailored to their unique needs, according to Devgan.

“It’s a domain specific product. It can do something a regular standard product cannot do,” he said, pointing to Tesla’s work on chips for Full Self Driving, and phone makers’ mobile computing devices that run all day on a single battery charge.

Like all companies dependent on components to power new products, the big players want to have assurance they can meet schedules, and an in-house chip design capability can help there, Devgan continued.

“You have some schedule, you want some control over that,” he told the JP Morgan conference attendees.

For the in-house design to work economically, scale of market is crucial. AI’s apparent boundless opportunity works for the hyperscalers here.

In the end, their in-house designed chip may cost less, when they cut the big chip maker’s over-size role out of the cost equation.

Where does this work? As always…”it depends.”

“It depends on each application, how much it costs, but definitely in AI there is volume, and volume is growing,” Devgan said, and he went on to cite mobile phones, laptops and autos as areas where the volume will drive the trend of custom chip creation.

Devgan declined to estimate how much system houses will take on the task of chip design going forward. Cadence wins in either case, by selling tools to semiconductor manufacturers, hyperscaling cloud leaders and  system houses.

He said: “We will leave that for the customer and the market to decide. Our job is to support both fully, and we are glad to do that.”

The trend bears watching. Years of technology progress has been based on system houses and their customers working with standard parts. Trends like in-house chip design may have the momentum to drastically rejigger today’s IT vendor Who’s Who, which has already been thoroughly rearranged in the wake of the cloud and the web. -jv

OpenAI GPT-4o Lands with Mini Thud or: Generative AI balances Hype and Reality in Chatbot Market Quest

May 19, 2024 By Jack Vaughan

It’s still too early to gauge Generative AI’s limits. That is another way of saying a circus atmosphere of hyperbole and demo theatrics is far from played out. The word “plateau” is heard today, and maybe a leveling off is only natural.

But now the uncertain space of ‘what it can’t do yet’ is mined each day. If Generative AI efforts plateau, and it merely changes the chatbot market as we know it, Generative AI will go down as a really big large-language disappointment.

This week’s OpenAI rollout of GPT-4o didn’t help. One can’t blame the OpenAI crew for trying their best to present awe inspiring on-stage demos as they saddle onto Danish Modern furniture set that bodes a comforting future. After all, there’s need to show their labs’ work is world changing or — barring that — fun.

The upstart’s was one among other episodes in the week’s AI Wars, as OpenAI’s demo was joined by Google I/O’s product roll outs on another stage in another corner of the Web, reported by Sean Michael Kerner and others.

For their part, the OpenAI crew walked through pet tricks, such as, asking the applications to translate “Why do we do linear equations?” into sparkling Italian.

Google’s show was just as breathless.

Yes, for OpenAI, the free app is a step into a new realm. (Although, as George Lawton points out, “free” is always an onion to unpeel.) And, yes, it vastly surpasses a voice-to-text demo of the 1990s. Does it move the bar much further than the Smart Speaker did in the mid-2010s? Let an army of pundits ponder this.

Our take: OpenAI’s announcement of something akin to a free-tier product was a bit short on awe. We’d second Sharon Goldman of Fortune who marked GPT-4o as “OpenAI’s emotive on steroids voice assistant.”

Of course, more accessible and easier entry for a wider range of people is OpenAI’s ticket to broadening into consumer markets. That’s where the killer app that justifies big valuations may be. Gain the consumer, and the enterprise follows.

That’s where OpenAI will meet the public and duke it out with friends like AWS, Google and Microsoft. There’s Apple too, which is likely now prepping spirited demos that show it has heard the bubbling cries of drowning users of Siri.

The next battle will be different than what has come before for heavily financed OpenAI. This stage in the technology’s evolution brings the OpenAI boffins down from the high ground, They say “hello whitebread-light demo patter” — just like Google, AWS or Microsoft product managers!

For a company that’s gained outsized attention in big headline deals for crucial infrastructure for big cloud players, it’s time to move toward apps. If it is to gain ground on a big scale, it will have to reach consumers. We take that as a less than nuanced theme in the GPT-4o roll out.

Cousin IoT: Brave New World Update

As if by chance, we sat in on Transforma’s report on IoT markets this very same week. While ably detailing the currents and eddies of IoT in the decades to come It seemed to convey a message relevant to Gen AI’s future course.

It’s been a long time since IoT first promised a brave new technology future — and such promises were never quite on the scale of Generative AI — but IoT has been grinding away gainfully, nevertheless.

IoT industry players have faced the same kind of existential challenge that GenAI is about to encounter. That is: The need to find a killer consumer app that it can power.

Transforma’s recent survey reports that there were 16.1 billion active IoT devices at the end of 2033. Annual device sales will grow from 4.1 billion in 2023 to 8.7 billion (a CAGR of 8%).

Yet, the world — even the industrial market within the bigger world — seems little changed. IoT’s top use cases, now and looking forward, fall short in terms of the energetic dynamism represented in early visions of IoT that looked more like StarTrek or the Jetsons.

Transforma looking toward the top three IoT use cases in 2033 cited 1- electronic labels; 2-building lights; and 3-headphones. You can come knocking because the van is not rocking, at least in terms of excitement. Still, these use cases represent real businesses.

Now, the assertion here — that Generative AI will be viewed in the future much as IoT is viewed today — is tentative. The agreement here is likely inexact … but may be useful for predicasts. Finally, my purpose here is not to put-down these young technologists’ efforts, but just to suggest that OpenAI and underlying Generative AI are in for a tough fight. — Jack Vaughan

Observer Monitor Dispatch: Dynatrace, StarTree, More

May 10, 2024 By Jack Vaughan

Kubernetes defense in focus as Dynatrace rolls with Runecast tech

Kubernetes’ quick rise to prominence in cloud computing may have left a few holes in applications’ defenses. That is something Dynatrace looks to address with Kubernetes Security Posture Management (KSPM) software. It’s said to employ observability data to enable quick response and mitigation of risks.

Dynatrace has a lineage when it comes to AI, originally arising out of the movement that placed AI agents on network nodes in order to track activity. KSPM employs  Dynatrace’s Davis hypermodal AI which combines predictive AI, Causal AI and Generative AI methods. The company said KSPM, thus accoutered, can ably convey immediate context for decision making as threats occur.

The company said  KSPM follows the integration of Runecast cloud native technology into the Dynatrace platform following the company’s successful acquisition earlier this year. Runecast technology supports continuous Kubernetes vulnerability scans, security compliance based on best practices, and remediation recommendations.

AI continues to find  renewed influence in the observability space. Updates are coming quickly, as this follow latest follows up on Dynatrace’s January roll-out of AI Observability extensions for large language models (LLMs).

 

StarTree Cloud extends observability

StarTree Cloud gains new observability and anomaly detection capabilities as well as vector search capabilities for its underlying Apache Pinot database engine with a new release reported by StarTree.

StarTree offers these services to customers of a cloud-based database-as-a-service that is specially targeting  analytics jobs. Like Confluent, StarTree arose out of the open-source activity of LinkedIn during the 2010s.

While Confluent has focused most effectively on data ingestion, StarTree has concentrated on data analytics, based on an implementation of the Pinot OLAP distributed columnar database.

The StarTree product suite is said to serve user-facing applications where a broad user base can query real-time data. The company noted DoorDash as a customer in this regard. The company said it partners with cloud and big data players such as AWS, Google Cloud, Microsoft, Confluent, Databricks and others in customer engagements.

The observability functionality new to the platform should allow  StarTree users, serving as developers or system administrators, to troubleshoot issues that arise within their cloud-based applications.

StarTree announced the general availability of StarTree ThirdEye software, offering multidimensional anomaly detection. As well, a write API supporting real-time sync for ELT pipelines such as Debezium, Fivetran, or dbt , is now in “private preview,” and integrations with visualization platforms, including Tableau and Grafana are available now.

Like others, StarTree is bringing nearest-neighbor vector search capabilities to its users, as well as users of the open-source Apache Pinot project.

Besides DoorDash, StarTree cites customers such as  Citi, Stripe, Nubank, and Zomato. For its part, in March, Citi announced a strategic investment in StarTree.

At the time,  Katya Chupryna, Director, Markets Strategic Investments at Citi marked StarTree and the underlying Pinot engine for speed of data ingestion.

“User-facing analytics have seen profound growth in recent years across all industries, accelerating the need for enterprise-ready, real-time data solutions. StarTree and Pinot’s speed of ingestion is unmatched on complex queries over rapidly changing data,” Chupruna said.

Also on tap

Cisco announced a virtual appliance for its AppDynamics On-Premises application observability offering. It’s aimed at users looking for customer-managed observability for on-premises deployments or cloud-based deployments where the customer retains control of all data and associated operations. AI-Powered Detection and Remediation with Cisco’s Cognition Engine is said to speed anomaly detection and root cause analysis … Riverbed announced Riverbed Unified Agent which allows IT to add SaaS-delivered visibility modules – for example, for end user experience and network monitoring – without adding more agents. Riverbed’s Platform initially launches with approximately 35 pre-built application integrations for third-parties including ServiceNow, Dynatrace, AppDynamics and DataDog. A Topology Viewer generates dynamic mapping of connected devices, while Riverbed NPM+, using the Riverbed Unified Agent, is said to overcome network blind spots created by remote work, public cloud, and encrypted architectures such as Zero Trust environments. This, while extending packet visibility to collected decrypted data at every user and server endpoint, including gaps such as encrypted tunnels in Zero Trust architectures. [PG]

Progressive Gauge

Copyright © 2025 · Jack Vaughan · Log in

  • Home
  • About
  • Blog
  • Projects, Samples and Items
  • Video
  • Contact Jack