Error message

  • Warning: date_timezone_set() expects parameter 1 to be DateTime, boolean given in format_date() (line 2062 of /home/scslive/public_html/includes/
  • Warning: date_format() expects parameter 1 to be DateTimeInterface, boolean given in format_date() (line 2072 of /home/scslive/public_html/includes/
  • Warning: date_timezone_set() expects parameter 1 to be DateTime, boolean given in format_date() (line 2062 of /home/scslive/public_html/includes/
  • Warning: date_format() expects parameter 1 to be DateTimeInterface, boolean given in format_date() (line 2072 of /home/scslive/public_html/includes/
  • Warning: date_timezone_set() expects parameter 1 to be DateTime, boolean given in format_date() (line 2062 of /home/scslive/public_html/includes/
  • Warning: date_format() expects parameter 1 to be DateTimeInterface, boolean given in format_date() (line 2072 of /home/scslive/public_html/includes/
  • Warning: date_timezone_set() expects parameter 1 to be DateTime, boolean given in format_date() (line 2062 of /home/scslive/public_html/includes/
  • Warning: date_format() expects parameter 1 to be DateTimeInterface, boolean given in format_date() (line 2072 of /home/scslive/public_html/includes/

Your Guide to Cisco Live 2018 in Orlando, Florida!

Theresa Miller - Tue, 06/12/2018 - 05:30

This week is the annual Cisco Live conference, where the attendees are spending the time with their favorite Disney and Universal characters in Orlando, Florida!  Beyond the superheroes and animated characters, let’s take a look at what this conference has for technical attendees, and what opportunities the conference provides. First of all, the conference is […]

The post Your Guide to Cisco Live 2018 in Orlando, Florida! appeared first on 24x7ITConnection.

A great 21st Citrix Synergy for Liquidware (sponsored)

Citrix UK User Group - Thu, 05/31/2018 - 21:20

Citrix Synergy 2018 took place in Anaheim, California May 8-10th. I was at the first Citrix “Thinergy” in 1998 at the Swan Hotel in Orlando. The conference remains one of the most Windows desktop focused conferences in the industry. The …

Read more »

The post A great 21st Citrix Synergy for Liquidware (sponsored) appeared first on UK Citrix User Group.

Everything to the Cloud! How Uila can help with your Pre-migration Assessment

Theresa Miller - Thu, 05/31/2018 - 05:30

Whether your organization will be moving everything to the cloud or not, most organizations are looking at what it means to migrate at least some of their applications or workloads.  Interdependencies may force organizations to run in a hybrid configuration where some infrastructure is on-premises, but a majority of the workload is now running in […]

The post Everything to the Cloud! How Uila can help with your Pre-migration Assessment appeared first on 24x7ITConnection.

Inexpensive GPU Virtualization Options for Testing

Helge Klein - Fri, 05/25/2018 - 00:01

For a new chapter in my ongoing series of talks and articles about browser performance I wanted to examine how key user experience metrics like page load time are dependent on available hardware resources. I prepared a VM and started my tests with a single CPU core and 2 GB of RAM, planning to gradually add more cores, more RAM and, eventually, a GPU. As it turned out, the latter is harder than anticipated.

Virtualizing a GPU – the Options

There seem to be many different ways to equip a VM with a GPU these days, but not all of them make sense for a test environment. Nvidia Grid, specifically, is pretty expensive, given that it requires specific professional GPUs and supported server hardware, unlocked by Grid software licenses.

Another variant that I did not find suitable: GPU-enabled VMs, e.g. from AWS or Azure. The reason being not even cost (I would only have used the VM for a dozen hours or so), but flexibility (I had to be able to change the VM’s hardware configuration).

With Nvidia Grid and a cloud VM out of the way, I was left with the following options. Please note that I was looking for a simple and inexpensive solution.

  1. Client Hyper-V with RemoteFX
  2. Client Hyper-V with direct device assignment (DDA)
  3. VMware Workstation

Let’s take a look at each of these in turn.

Client Hyper-V with RemoteFX Configuration

RemoteFX GPU virtualization has been available in Windows 10 Client Hyper-V since version 1511. Some docs state that the Enterprise SKU of Windows 10 is required in the guest, but I did not find that to be correct: Pro and Enterprise gave identical results.

To enable RemoteFX GPU virtualization:

  • Host: in Hyper-V settings, enable a physical GPU for use with RemoteFX
  • Guest: in the VM’s settings, add a RemoteFX 3D Video Adapter to the VM

No need to care about GPU drivers in the VM: they are part of Hyper-V Integration Services.


With the above steps completed, you will see a virtual GPU in the VM (Microsoft RemoteFX Graphics Device – WDDM):

DXDiag seems to be quite happy with it:

Chrome not so much:

Also, there is no GPU in Task Manager:


RemoteFX does not map (part of) the physical GPU into the VM. Instead, it virtualizes certain DirectX APIs. That may be enough for some uses cases, but Chrome is looking for something different and consequently sticks to software rendering.

By the way, a good indicator that there is no “real” GPU available in the VM is that you do not need to install GPU vendor drivers.

Client Hyper-V with Direct Device Assignment (DDA) DDA Requirements

DDA maps a physical GPU into a VM. While it is documented only for Windows Server I thought it just might work with Windows 10, too.

With DDA, the VM “sees” a real GPU. Therefore vendor drivers need to be installed in the VM.

DDA only works if the host machine’s BIOS supports SR-IOV. In other words, the BIOS must support relinquishing control of the PCI Express bus to the OS. That, unfortunately, is a capability typically only found in server hardware. SR-IOV is not to be confused with VT-d, which is typically found in consumer hardware but which is not to sufficient to make DDA work.

DDA and Licensing

Even if you have a machine that meets the hardware requirements for DDA outlined above, you may run into a problem of a very different nature: Nvidia’s consumer cards cannot be used because of restrictions in the driver. Nvidia’s Grid cards, on the other hand, require software licenses in addition to the professional-grade hardware. Rachel Berry has a good overview of Nvidia Grid licensing.

AMD GPUs (purportedly) do support DDA, but I did not test that.

VMware Workstation Configuration

The first thing you may notice after installing VMware Workstation is that it does not work if Hyper-V is enabled. If you want to keep using Hyper-V your best option is to create a new boot menu entry with Hyper-V disabled. That way you can switch between Hyper-V and VMware Workstation simply by rebooting. From Stack Overflow:

C:\>bcdedit /copy {current} /d "No Hyper-V" The entry was successfully copied to {ff-23-113-824e-5c5144ea}. C:\>bcdedit /set {ff-23-113-824e-5c5144ea} hypervisorlaunchtype off The operation completed successfully.

As with RemoteFX, there is no need to care about GPU drivers in the VM: they are part of VMware Tools. This indicates that the concept is similar in nature to RemoteFX.


The virtual GPU you will see in the VM is of the type VMware SVGA 3D:

Again, DXDiag is happy:

Again, Chrome not so much:

As with RemoteFX, there is no GPU in Task Manager:


From an application’s point of view, VMware Workstation GPU virtualization is similar to RemoteFX: potentially good enough for some use cases, but there is no “real” GPU in the VM.


It’s 2018. This should be easier!

GPU virtualization is most definitely not mainstream yet. Currently, server hardware with either AMD or Nvidia Grid GPUs is required. That makes testing unnecessarily difficult and expensive.

The post Inexpensive GPU Virtualization Options for Testing appeared first on Helge Klein.

Microsoft has announced a few changes to the future of the Office suite

Theresa Miller - Tue, 05/22/2018 - 05:30

Microsoft has announced a few changes to the future of the Office suite that most IT people will need to understand for the future of the environment they manage. If you’re looking after a decent sized IT environment, then you’re probably also looking after Microsoft Office, and it’s probably one of the most important apps […]

The post Microsoft has announced a few changes to the future of the Office suite appeared first on 24x7ITConnection.

Review of Additive Manufacture and Generative Design for PLM/Design at Develop 3D Live 2018

Rachel Berrys Virtually Visual blog - Wed, 05/16/2018 - 13:54

A couple of months ago, back at D3DLive! I had the pleasure of chairing the Additive Manufacturing (AM) track. This event in my opinion alongside a few others e.g. Siggraph and COFES is one of the key technology and futures events for the CAD/Graphics ecosystem. This event is also free thanks in part to major sponsors HP, Intel, AMD and Dell sponsorship.

A few years ago, at such events the 3D-printing offerings were interesting, quirky but not really mainstream manufacturing or CAD. There were 3D-printing vendors and a few niche consultancies, but it certainly wasn’t technology making keynotes or mentioned by the CAD/design software giants. This year saw the second session of the day on the keynote stage (video here) featuring a generative design demo from Bradley Rothenberg of nTopology.

With a full track dedicated to Additive Manufacture(AM) this year including the large mainstream CAD software vendors such as Dassault, Siemens PLM and Autodesk this technology really has hit the mainstream. The track was well attended with approximately half of the attendees when poled where actually involved in implementing additive manufacture and a significant proportion using it in production.

There was in general a significant overlap between many of the sessions, this technology has now become so mainstream that rather than seeing new concepts we are seeing like mainstream CAD more of an emphasis on specific product implementations and GUIs.

The morning session was kicked off by Sophie Jones, General Manager of Added Scientific a specialist consultancy with strong academic research links who investigate future technologies. This really was futures stuff rather than the mainstream covering 3D-printing of tailored pharmaceuticals and healthcare electronics.

Kieron Salter from KWSP then talked about some of their user case studies, as a specialist consultancy they’ve been needed by some customers to bridge the gaps in understanding. In particular, some of their work in the Motorsports sector was particularly interesting as cutting-edge novel automotive design.

Jesse Blankenship from Frustum gave a nice overview of their products and their integration into Solid Edge, Siemens NX and Onshape but he also showed the developer tools and GUIs that other CAD vendors and third-parties can use to integrate generative design technologies. In the world of CAD components, Frustum look well-placed to become a key component vendor.

Andy Roberts from Desktop Metal gave a rather beautiful demonstration walking through the generative design of a part, literally watching the iteration from a few constraints to an optimised part. This highlighted how different many of these parts can be compared to traditional techniques.

The afternoon’s schedule started with a bonus session that hadn’t made the printed schedule from Johannes Mann of Volume Graphics. It was a very insightful overview of the challenges in fidelity checking additive manufacturing and simulations on such parts (including some from Airbus).

Bradley Rothenberg of nTopology reappeared to elaborate on his keynote demo and covered some of the issues for quality control and simulation for generative design that CAM/CAE have solved for conventional manufacturing techniques.

Autodesk’s Andy Harris’ talk focused on how AM was enabling new genres of parts that simply aren’t feasible via other techniques. The complexity and quality of some of the resulting parts were impressive and often incredibly beautiful.

Dassault’s session was given by a last-minute speaker substitution of David Reid; I haven’t seen David talk before and he’s a great speaker. It was great to see a session led from the Simulia side of Dassault and how their AM technology integrates with their wider products. A case study on Airbus’ choice and usage of Simulia was particularly interesting as it covered how even the most safety critical, traditional big manufacturers are taking AM seriously and successfully integrating it into their complex PLM and regulatory frameworks.

The final session of the day was probably my personal favourite, Louise Geekie from Croft AM gave a brilliant talk on metal AM but what made it for me was her theme of understanding when you shouldn’t use AM and it’s limitations – basically just because you can… should you? This covered long term considerations on production volumes, compromises on material yield for surface quality, failure rates and costs of post-production finishing. Just because a part has been designed by engineering optimisation doesn’t mean an end user finds it aesthetically appealing – the case where a motorcycle manufacturer and indeed wants the front fork to “look” solid.

Overall my key takeaways were:

·       Just because you can doesn’t mean you should, choosing AM requires an understanding of the limitations and compromises and an overall plan if volume manufacture is an issue

·       The big CAD players are involved but there’s still work to be done to harden the surrounding frameworks in particular reliable simulation, search, fidelity testing.

·       How well the surrounding products and technologies handle the types of topologies and geometries GM throws out will be interesting. In particular it’ll be interesting to watch how Siemens Syncronous Technology and direct modellers cope, and the part search engines such as Siemens Geolus too.

·       Generative manufacture is computationally heavy and the quality of your CPU and GPU is worth thinking about.

Hardware OEMS and CPU/GPU Vendors taking CAD/PLM seriously

These new technologies are all hardware and computationally demanding compared to the modelling kernels of 20 years ago. AMD were showcasing and talking about all the pro-viz, rendering and cloud graphics technologies you’d expect but it was pleasing to see their product and solution teams and those from Dell, Intel, HP etc talking about computationally intensive technologies that benefit from GPU and CPU horse power such as CAE/FEA and of course generative design. It’s been noticeable in recent years in the increasing involvement and support from hardware OEMs and GPU vendors for end-user and ISV CAD/Design events and forums such as COFES, Siemens PLM Community and Dassault’s Community of Experts; which should hopefully bode well for future platform developments in hardware for CAD/Design.


A few weeks ago Al Dean from Develop3D wrote an article (bordering on a rant) about how poorly positioned a lot of the information around generative design (topology optimisation) and it’s link to additive manufacture is. I think many reading, simply thought – yes!

After reading it – I came to the conclusion that many think generative design and additive manufacture are inextricably linked. Whilst they can be used in conjunction there are vast numbers of use cases where the use of only one of the technologies is appropriate.

Generative design in my mind is computationally optimising a design to some physical constraints – it could be mass of material, or physical forces (stress/strain) and could include additional constraints – must have a connector like this in this area, must be this long or even must be tapered and constructed so it can be moulded (include appropriate tapers etc – so falls out the mold).

Additive manufacture is essentially 3-D printing, often metals. Adding material rather than the traditional machining mentality of CAD (Booleans often described as target and tool) – removing stuff from a block of metal by machining.

My feeling is generative design far greater potential for reducing costs and optimising parts for traditional manufacturing techniques e.g. 3/5-axis G-code like considerations, machining, injection molding than has been highlighted. Whilst AM as a prototyping workflow for those techniques is less mature than it could be as the focus has been on these weird and wonderful organic parts you couldn’t make before without AM/3-D Printing.

Microsoft Build’s hidden IT Pro announcements

Theresa Miller - Tue, 05/15/2018 - 05:30

Microsoft Build is the annual gathering of the software developer community who are focused on the Microsoft platform, held recently. An event to rival Microsoft Ignite, it also is a chance for Microsoft to make some big product announcements. For this reason, it’s worth it for IT Pros to also look at the event coverage […]

The post Microsoft Build’s hidden IT Pro announcements appeared first on 24x7ITConnection.

Three New vSphere 6.7 Features That Make the Upgrade Worth It

Theresa Miller - Thu, 05/10/2018 - 05:30

VMware has a long history of innovation, with many waiting with baited breath for new vSphere releases. From Storage vMotion to the VCSA, there has been so many interesting features released over the years. Each release alone is packed full of new features that can sometimes be hard to sort through, which is why I […]

The post Three New vSphere 6.7 Features That Make the Upgrade Worth It appeared first on 24x7ITConnection.

Citrix Synergy 2018 Keynote Highlights just for you!

Theresa Miller - Tue, 05/08/2018 - 06:15

This morning we all work up to a mild earthquake to get the day started before the event today.  Thankfully just a mild rumble that quickly passed.  The energy at Citrix Synergy 2018 brings promise of a great event, and some new products to look forward to.  At the keynote there were many things uncovered, […]

The post Citrix Synergy 2018 Keynote Highlights just for you! appeared first on 24x7ITConnection.

Download, Install, Import Visual C++ Redistributables with VcRedist

Aaron Parker's stealthpuppy - Mon, 05/07/2018 - 21:32

Last year I wrote a PowerShell script that can download, install or import the Visual C++ Redistributables into MDT or ConfigMgr. Long-term maintenance of the full feature set in a single script is a little unwieldy so I’ve re-written the script and created a PowerShell module – VcRedist.

Refactoring the script into a module has been a great little project for creating my first PowerShell function and publishing it to the PowerShell Gallery.

Why VcRedist?

At this point, I’m sure you’re saying to yourself – “Aaron, haven’t you just created Chocolatey?”. In a way yes, this module does exactly what you can do with Chocolatey – install the Visual C++ Redistributables directly to the local machine. Although you can download and install all of the supported (and unsupported) Redistributables, the primary aim of the module is to provide a fast way to download and import the Redistributables into the Microsoft Deployment Toolkit or System Center Configuration Manager for operating system deployments.


The VcRedist module is published to the PowerShell Gallery, which means that it’s simple to install the module and starting importing with a few lines of PowerShell. For example, here’s how you could install the module, download all of the supported Redistributables and import them into an MDT deployment share:

Install-Module -Name VcRedist Import-Module VcRedist $VcList = Get-VcList | Get-VcRedist -Path "C:\Temp\VcRedist" Import-VcMdtApp -VcList $VcList -Path "C:\Temp\VcRedist" -MdtPath "\\server\share\Reference"

This results in each of the Visual C++ Redistributables imported as a separate application with all necessary properties including Version, silent command line, Uninstall Key and 32-bit or 64-bot operating system support.

Visual C++ Redistributables imported into an MDT share with VcRedist

The same approach can be used to import the Redistributables into a ConfigMgr site:

Install-Module VcRedist Import-Module VcRedist $VcList = Get-VcList | Get-VcRedist -Path "C:\Temp\VcRedist" Import-VcCmApp -VcList $VcList -Path "C:\Temp\VcRedist" -CMPath "\\server\share\VcRedist" -SMSSiteCode LAB

Just like MDT, each Redistributable is imported into ConfigMgr; however, Import-VcCmApp copies the Redistributables to a share for distribution and creates and application with a single deployment for each one.

Visual C++ Redistributables imported into ConfigMgr with VcRedist

Of course, the module can download and install the Redistributables to the local machine:

Install-Module VcRedist Import-Module VcRedist $VcList = Get-VcList | Get-VcRedist -Path "C:\Temp\VcRedist" $VcList | Install-VcRedist -Path C:\Temp\VcRedist

By default, this installs all of the supported Redistributables:

Visual C++ Redistributables installed locally with VcRedist

Note that the 2015 and 2017 Redistributables are the same version, so the end result will include only the 2017 versions.


This module includes the following functions:


This function reads the Visual C++ Redistributables listed in an internal manifest or an external XML file into an array that can be passed to other VcRedist functions. Running Get-VcList will return the supported list of Visual C++ Redistributables. The function can read an external XML file that defines a custom list of Visual C++ Redistributables.


Run Export-VcXml to export the internal Visual C++ Redistributables manifest to an external XML file. Use -Path to define the path to the external XML file that the manifest will be saved to. By default Export-VcXml will export only the supported Visual C++ Redistributables.


To download the Visual C++ Redistributables to a local folder, use Get-VcRedist. This will read the array of Visual C++ Redistributables returned from Get-VcList and download each one to a local folder specified in -Path. Visual C++ Redistributables can be filtered for release and processor architecture.


To install the Visual C++ Redistributables on the local machine, use Install-VcRedist. This function again accepts the array of Visual C++ Redistributables passed from Get-VcList and installs the Visual C++ Redistributables downloaded to a local path with Get-VcRedist. Visual C++ Redistributables can be filtered for release and processor architecture.


To install the Visual C++ Redistributables as a part of a reference image or for use with a deployment solution based on the Microsoft Deployment Toolkit, Import-VcMdtApp will import each of the Visual C++ Redistributables as a separate application that includes silent command lines, platform support and the UninstallKey for detecting whether the Visual C++ Redistributable is already installed. Visual C++ Redistributables can be filtered for release and processor architecture.

Each Redistributables will be imported into the deployment share with application properties for a successful deployment.


To install the Visual C++ Redistributables with System Center Configuration Manager, Import-VcCmApp will import each of the Visual C++ Redistributables as a separate application that includes the application and a single deployment type. Visual C++ Redistributables can be filtered for release and processor architecture.

Tested On

Tested on Windows 10 and Windows Server 2016 with PowerShell 5.1. Install-VcRedist and Import-VcMdtApp require Windows and the MDT Workbench. Get-VcList, Export-VcXml and Get-VcRedist do work on PowerShell Core; however, most testing is completed on Windows PowerShell.

To Do

Right now, I have a few tasks for updating the module, including:

  • Additional testing / Pester tests
  • Add -Bundle to Import-VcMdtApp to create an Application Bundle and simplify installing the Redistributables
  • Documentation updates

For full details and further updates, keep an eye on the repository and test out the module via the PowerShell Gallery.

Image credit:

Alexey Ruban

This article by Aaron Parker, Download, Install, Import Visual C++ Redistributables with VcRedist appeared first on Aaron Parker.

Categories: Community, Virtualisation

Examine information sources before you share

Theresa Miller - Sat, 05/05/2018 - 16:48

Humans transmit information using language, but how often do you stop and examine information sources for the content you consume? Originally, humans shared stories to transmit this information. Then our scholars and holy men began to write the information down. Eventually, thanks to the printing press, information that was written down could be produced and […]

The post Examine information sources before you share appeared first on 24x7ITConnection.

Deploying a custom OMS Log Analytics Workspace via GitHub – Avoid problems with ARM templates

Nicholas Dille (Sepago) - Thu, 05/03/2018 - 19:58
Azure is “my” cloud with a lot of platform services allowing user, programmers and DevOps building powerful and scalable solutions. One of my favorite ones is Azure OMS Log Analytics – a big data platform with a great query language and professional dashboards. In the past I...
Categories: , Citrix, Virtualisation

AWS and NICE DVC – a happy marriage! … resulting in a free protocol on AWS

Rachel Berrys Virtually Visual blog - Thu, 05/03/2018 - 13:12

It’s now two years since Amazon bought NICE and their DVC and EnginFrame products. NICE were very good at what they did. For a long time they were one of the few vendors who could offer a decent VDI solution that supported Linux VMs, with a history in HPC and Linux they truly understood virtualisation and compute as well as graphics. They’d also developed their own remoting protocol akin to Citrix’s ICA/HDX and it was one of the first to leverage GPUs for tasks like H.264 encode.

Because they did Linux VMs and neither Citrix nor VMware did, NICE were often a complementary partner rather than a competitor although with both Citrix and VMware adding Linux support that has shifted a little. AWS promised to leave NICE DVC products alone and have been true to that. However the fact Amazon now owns one of the best and experience protocol teams around has always raised the possibility they could do something a bit more interesting than most other clouds.

Just before Xmas in December 2017 without much fuss or publicity, Amazon announced that they’d throw NICE DVC in for free on AWS instances.

NICE DCV is a well-proven product with standalone customers and for many users offers an alternative to Citrix/VMware offerings; which raises the question why run VMware/Citrix on AWS if NICE will do?

There are also an awful lot of ISVs looking to offer cloud-based services and products including many with high graphical demands. To run these applications well in the cloud you need a decent protocol, some have developed their own which tend to be fairly basic H.264, others have bought in technology from the likes of Colorado Code Craft or Teradici’s standalone Cloud Access Software based around the PCoIP protocol. Throwing in a free protocol removes the need to license a third-party such as Teradici, which means the overall solution cost is cut but with no impact on the price AWS get for an instance. This could be a significant driver for ISVs and end-users to choose AWS above competitors.

Owning and controlling a protocol was a smart move on Amazon’s part, a key element of remoting and the performance of a cloud solution, it makes perfect sense to own one. Microsoft and hence Azure already have RDS/RDP under their control. Will we see moves from Google or Huawei in this area?

One niggle is that many users need not just a protocol but a broker, at the moment Teradici and many do not offer one themselves and users need to go to another third-party such as Leostream to get the functionality to spin-up and manage the VMs. Leostream have made a nice little niche supporting a wide range of protocols. It turns out that AWS are also offering a broker via the NICE EnginFrame technologies, this is however an additional paid for component but the single vendor offering may well appeal. It was really hard to find this out, I had to contact the AWS product managers for NICE to be certain. I really couldn’t work out what was available from the documentation and product overviews from AWS (in the end I had to contact the product management team directly).

Teradici do have a broker in-development, the details of which they discussed with Jack on

So, today there is the option of a free protocol and paid for broker (NICE+EngineFrame alibi tied to AWS) and soon there will be a paid protocol from Teradici with a broker thrown in, the protocol is already available on the AWS marketplace.

This is just one example of many where cloud providers can take functionality in-house and boost their appeal by cutting out VDI, broker or protocol vendors. For those niche protocol and broker vendors they will need to offer value through platform independence and any-ness (the ability to choose AWS, Azure, Google Cloud) against out of the box one-stop cloud giant offerings. Some will probably succeed but a few may well be squeezed. It may indeed push some to widen their offerings e.g. protocol vendors adding basic broker capabilities (as we are seeing with Teradici) or widening Linux support to match the strong NICE offering.

In particular broker vendor Leostream may be pushed, as other protocol vendors may well follow Teradici’s lead. However, analysts such as Gabe Knuth have reported for many years on Leostream’s ability to evolve and add value.

We’ve seen so many acquisitions in VDI/Cloud where a good small company gets consumed by a giant and eventually fails, the successful product dropped and the technologies never adopted by the mainstream business. AWS seem to have achieved the opposite with NICE, continuing to invest in a successful team and product whilst leeraging exactly what they do best. What a nice change! It’s also good to see a bit more innovation and competition in the protocol and broker space.

Dell Technologies World – On-Site Report

Theresa Miller - Wed, 05/02/2018 - 03:54

I’m writing this on Day 2 of Dell Technologies World 2018. The conference in this timeframe in years past was EMC World. When Dell and EMC merged, the show was named Dell EMC World. This year, the name was changed to Dell Technologies World to reflect the companies that are part of the Dell world. […]

The post Dell Technologies World – On-Site Report appeared first on 24x7ITConnection.

Migrating local VM owner certificates for VMs with vTPM

Microsoft Virtualisation Blog - Thu, 12/14/2017 - 06:27

Whenever I want to replace or reinstall a system which is used to run virtual machines with a virtual trusted platform module (vTPM), I’ve been facing a challenge: For hosts that are not part of a guarded fabric, the new system does need to be authorized to run the VM.
Some time ago, I wrote a blog post focused on running VMs with a vTPM on additional hosts, but the approach highlighted there does not solve everything when the original host is decommissioned. The VMs can be started on the new host, but without the original owner certificates, you cannot change the list of allowed guardians anymore.

This blog post shows a way to export the information needed from the source host and import it on a destination host. Please note that this technique only works for local mode and not for a host that is part of a guarded fabric. You can check whether your host runs in local mode by running Get-HgsClientConfiguration. The property Mode should list Local as a value.

Exporting the default owner from the source host

The following script exports the necessary information of the default owner (“UntrustedGuardian“) on a host that is configured using local mode. When running the script on the source host, two certificates are exported: a signing certificate and an encryption certificate.

Importing the UntrustedGuardian on the new host

On the destination host, the following snippet creates a new guardian using the certificates that have been exported in the previous step.

Please note that importing the “UntrustedGuardian” on the new host has to be done before creating new VMs with a vTPM on this host — otherwise a new guardian with the same name will already be present and the creation with the PowerShell snippet above will fail.

With these two steps, you should be able to migrate all the necessary bits to keep your VMs with vTPM running in your dev/test environment. This approach can also be used to back up your owner certificates, depending on how these certificates have been created.

Categories: Microsoft, Virtualisation


Subscribe to aggregator - Virtualisation