Archive for the Category
‘From: Network World’

Cyber security remains a hot topic with nearly every IT and business leader that I speak with. In particular, there seems to be an intensified focus on network security. Security is typically deployed in layers (network, compute and application), and I expect that model to continue in the short-term, but given the fact that many of the building blocks of digitization, such as IoT and the cloud, are network-centric, there should be a stronger focus on leveraging the network and network-based security to protect the organization.

Star Trek is filled with advice that apply to today’s tech professionals. Here’s a look at seven from the Enterprise’s most logical crew member, Spock.

It’s no surprise that many network engineers are also fans of Star Trek. Personally, I have been a Trekkie for as long as I can remember. One of the appealing things about Star Trek is that it pushed the limits of what’s possible. In fact, many technologies we take for granted today were previewed on Star Trek over 50 years ago. Things such as wireless communications, immersive videoconferencing and tablet computers were all used regularly on the Starship Enterprise long before we used them down on Earth.

Nvidia’s TensorRT 3 optimizes and compiles complex networks to get the best possible performance for AI inferencing.

It’s safe to say the Internet of Things (IoT) era has arrived, as we live in a world where things are being connected at pace never seen before. Cars, video cameras, parking meters, building facilities and anything else one can think of are being connected to the internet, generating massive quantities of data.

The question is how does one interpret the data and understand what it means? Clearly trying to process this much data manually doesn’t work, which is why most of the web-scale companies have embraced artificial intelligence (AI) as a way to create new services that can leverage the data. This includes speech recognition, natural language processing, real-time translation, predictive services and contextual recommendations. Every major cloud provider and many large enterprises have AI initiatives underway.

However, many data centers aren’t outfitted with enough processing power for AI inferencing. For those not familiar with the different phases of AI, training is teaching the AI new capabilities from an existing set of data. Inferencing is applying that learning to new data sets. Facebook’s image recognition and Amazon’s recommendation engine are both good examples of inferencing.

This week at its GPU Technology Conference (GTC) in China, Nvidia announced TensorRT 3, which promises to improve the performance and cut the cost of inferencing. TensorRT 3 takes very complex networks and optimizes and compiles them to get the best possible performance for AI inferencing. The below graphic shows that it acts as AI “middleware” so the data can be run through any framework and sent to any GPU. Recall this post where I explained why GPUs were much better for AI applications than CPUs. Nvidia has a wide range of GPUs, depending on the type of application and processing power required.

Unlike other GPU vendors, Nvidia’s approach isn’t just great silicon. Instead it takes an architectural approach where it combines software, development tools and hardware as an end-to-end solution.

During his keynote, CEO Jensen Huang showed some stats where TensorRT 3 running on Nvidia GPUs offered performance that was 150x better than CPU-based systems for translation and 40x better for images, which will save its customer huge amounts of money and offer a better quality of service. I have no way of proving or disproving those numbers, but I suspect they’re accurate because no other vendor has the combination of a high-performance compiler, run-time engine and GPU optimized to work together.

Other Nvidia announcements

  • DeepStream SDK introduced. It delivers low-latency video analytics in real time. Video inferencing has become a key part of smart cities but is being used in entertainment, retail and other industries as well.
  • An upgrade to CUDA, Nvidia’s accelerated computing software platform. Version 9 is now optimized for the new Tesla V100 GPU accelerators, which is the highest-end GPU and ideal for AI, HPC and graphically intense applications such as virtual reality.
  • Huawei, Inspur and Lenovo using Nvidia’s HGX reference architecture to offer Volta-based systems. The server manufacturers will be granted early access to HGX architectures for data centers and design guidelines. The HGX architecture is the same one used by Microsoft and Facebook today, meaning Asia-Pac-based organizations can have access to the same GPU-based servers as the leading web-scale cloud providers.

The world is changing quickly, and it’s my belief that market leaders will be defined by the organizations that have the most data and the technologies to interpret that data. Core to that is GPU-based machine learning and AI, as these systems can do things far faster than people.

As someone who has been following enterprise WAN architectures for decades, I find their evolution fascinating, especially the number of new technologies that have been deployed in isolation. For example, WAN optimization and SD-WANs are often discussed as separate solutions.  From my perspective, I can’t fathom why a business would deploy an SD-WAN and not implement WAN optimization as part of it.  If you’re going to go through the work of modernizing your WAN architecture, then why wouldn’t you integrate optimization technologies into your deployment right from the start?

The future of data centers will rely on cloud, hyperconverged infrastructure and more powerful components

A data center is a physical facility that enterprises use to house their business-critical applications and information, so as they evolve, it’s important to think long-term about how to maintain their reliability and security.

Data center components

Data centers are often referred to as a singular thing, but in actuality they are composed of a number of technical elements such as routers, switches, security devices, storage systems, servers, application delivery controllers and more. These are the components that IT needs to store and manage the most critical systems that are vital to the continuous operations of a company. Because of this, the reliability, efficiency, security and constant evolution of a data center are typically a top priority.

With Intersight, Cisco facilitates cloud management on its UCS and HyperFlex HCI system, with potential use on its other networking products.

I don’t think anyone would argue with the premise that data centers have increased significantly over the past decade. Data centers used to be orderly, as each application had its own dedicated hardware and software. This was highly inefficient, but most data centers could be managed with a handful of people.

Then something changed. Businesses were driven to improve the utilization of infrastructure and increase the level of agility, and along came a number of technologies such as virtualization, containers and the cloud. Also, organizations started to embrace the concept of DevOps, which necessitates a level of dynamism and speed never seen before in data centers. 

Arista Any Cloud brings consistency to the cloud network and reduces complexity, so network professionals can deploy networks in hybrid clouds like they do in data centers.

Many years ago, when Arista Networks was in its infancy, its charismatic and sometimes controversial (at least to the folks at Cisco) CEO talked about how the company’s software-first approach would disrupt the networking industry. Just a few years later, the company stands a $1.7 billion revenue company with a dominant position in the webscale industry and a market cap of over $13 billion, so clearly CEO Jayshree Ullal’s prophecy came true.

Insight and Influence Through Social Media
ZK Research: Home
RSS Feed
ZK Research is proudly powered by WordPress | Entries (RSS) | Comments (RSS) | Custom Theme by The Website Taylor