All Posts By

LF Edge

SDxCentral: EdgeX Foundry Edinburgh Release Signals IoT Platform ‘Ready for Primetime’

By | EdgeX Foundry, In the News

EdgeX Foundry, an open source interoperable framework for edge IoT computing, dropped its fourth code release titled Edinburgh.

“We’re calling it an anchor release for commercial adoption across IoT use cases,” said Arpit Joshipura, general manager for Networking, Edge, and IoT at the Linux Foundation. At this point in the platform’s lifecycle, EdgeX community members have created a range of complementary products and services to support commercial deployments, he explained. This includes training and customer pilot programs, and plug-in enhancements for device connectivity, applications, data and system management, and security.

Read the SDxCentral article here.

The New Stack: How the Linux Foundation’s EVE Can Replace Windows, Linux for Edge Computing

By | In the News, Project EVE

Whether or not Edge computing serves as the backbone of mission-critical business worldwide depends on the success of the underlying network.

Recognizing the Edge’s potential and urgency to support Edge network, The Linux Foundation earlier this year created LF Edge, an umbrella organization dedicated to creating an open, agnostic and interoperable framework for edge computing. Similar to what the Cloud Native Computing Foundation (CNCF) has done for cloud development, LF Edge aims to enhance cooperation among key players so that the industry as a whole can advance more quickly.

By 2021, Gartner forecasts that there will be approximately 25 billion IoT devices in use around the world. Each of those devices, in turn, has the capacity to produce immense volumes of valuable data. Much of this data could be used to improve business-critical operations — but only if we’re able to analyze it in a timely and efficient manner. As mentioned above, it’s this combination of factors that has led to the rise of edge computing as one of the most rapidly -developing technology spaces today.

This idea of interoperability at the edge is particularly important because the hardware that makes up edge devices is so diverse — much more so than servers in a data center. Yet for edge computing to succeed, we need to be able to run applications right on local gateway devices to analyze and respond to IoT and Industry 4.0 data in near-real time. How do you design applications that are compatible with a huge variety of hardware and capable of running without a reliable cloud connection? This is the challenge that LF Edge is helping to solve.

Part of the solution is Project EVE, an Edge Virtualization Engine donated to LF Edge by ZEDEDA last month. I think of EVE as doing for the edge what Android did for mobile phones and what VMware did for data centers: decoupling software from hardware to make application development and deployment easier.

Read more at The News Stack here.

Using DockerHub in Akraino Edge Stack & Other Linux Foundation Projects

By | Akraino, Blog

By Kaly Xin, Eric Ball,  and Cristina Pauna

DockerHub is the world’s largest library and community for container images. It offers a huge repository for storing container images and it is available world wide. It can automatically build container images from GitHub and Bitbucket and push them to Docker Hub. These are just a few of the features it provides, but maybe one of the best features is that it offers seamless support for multi arch images through fat manifest.

Why Docker Hub is recommended for Multi-Arch

Docker Hub registry is able to store manifest list (or fat manifests). A manifest list acts as a pointer to other images built for a specific architecture thus making it possible to use the same name for images that are built on hardware with different architectures.

Figure 1: Docker registry storing amd64, arm64 images and their fat manifest

In the picture above akraino/validation:k8s-latest is the fat manifest, and its name can be used to reference both images akraino/validation:k8s-amd64-latest and akraino/validation:k8s-arm64-latest. Inspecting the manifest offers the details on what images it has, for what hardware architecture and what OS.

Figure 2: Docker fat manifest details

How does it work?

When building an image for a specific arch, the arch is added in the tag of the image (akraino/validation:k8s-amd64-latest and akraino/validation:k8s-arm64-latest).

After the images are pushed in the Docker Hub repo, the manifest can be created from the two images. Its name will be the same as the two images but with the arch removed from the tag (akraino/validation:k8s-latest).

To do this in CI with Jenkins, the Jenkins slave has to have docker and a couple of other LF tools installed. The connection to Docker Hub is done through LF scripts (see releng-global-jjb for more info) and all you need to do is define the jjb jobs .

The Akraino validation project is already pushing to Dockerhub, so if you would like to check out some template code, take a look at ci-management/jjb/validation. The docker images are pushed in the official repo and the docker build jobs are running daily. 

In the figure below, the main Jenkins job (validation-master-docker) triggers two parallel jobs that build and push into the Docker Hub registry the amd64 (akraino/validation:k8s-amd64-latest ) and arm64 (akraino/validation:k8s-arm64-latest ) images.  At the end, the fat manifest (akraino/validation:k8s-latest ) is done in a separate job. 

When pulling the image, the name of the manifest is used  (akraino/validation:k8s-latest); the correct image will be pulled based on the architecture of the host from which the pull is made.

Figure 4: Pulling a docker image from two different hardware architecture servers using the same name

What’s next

Docker Hub has been integrated in LF projects like OPNFV from the beginning and is now integrated in Akraino too, so other open source projects can refer to this successful experience to integrate Docker Hub in their pipeline.

What to Expect from the Open Glossary of Edge Computing in 2019

By | Blog, Open Glossary of Edge Computing

By Alex Marcham, LF Edge and State of the Edge Contributor and Technical Marketing Manager at Vapor IO, and Matt Trifiro, Co-Chair at State of the Edge; Chair at Open Glossary of Edge Computing and CMO at Vapor IO

The Open Glossary of Edge Computing began as a utilitarian appendix to the 2018 State of the Edge report. It had humble goals: to cut through the morass of vendor- and pundit-driven definitions around edge computing and, instead, deliver a crisply-defined common lexicon that would enhance understanding and accelerate conversations around all things edge.

Very quickly, it became clear that the Open Glossary could be a powerful and unifying force in the fledgling world of edge computing and that no one entity should own it. Instead, everybody should own it. Thus began our partnership with The Linux Foundation. We converted the glossary into a GitHub repo, placed it under a Creative Commons license, and created an open source project around it.

In January 2019, the Open Glossary became one of the founding projects of LF Edge, the Linux Foundation’s umbrella group for edge computing projects. The Open Glossary now plays a critical role that spans all of the LF Edge projects with its mission to collaborate around a single point of reference for edge computing terminology. This community-driven lexicon helps mitigate confusing marketing buzzwords by offering a foundation of clear and well-understood words and phrases that can be used by everyone.

In 2019, the momentum of the Open Glossary project will continue unabated. Open Glossary has four main projects this year, building on its successes in 2018:

Grow the Community

The Open Glossary project depends on open collaboration, and the community is actively building engagement by seeking out participation from key stakeholder groups in edge computing. Within The Linux Foundation itself, the Open Glossary team has sought input from not only all of the LF Edge projects but also adjacent projects, such as the CNCF’s Kubernetes IoT and Edge working group. In addition, the Open Glossary has formed alliances with other edge computing groups and foundations, including the TIA, iMasons, and the Open19 Foundation. Partnering with other non-profits and consortia will help the project grow its base of passionate collaborators who are dedicated to expanding and improving the Open Glossary.

The Taxonomy Project

Readers of the Open Glossary want more than mere definitions; they want to know how the constituent parts fit together as a whole. To answer this request, the Open Glossary has begun the Taxonomy Project, a working group that seeks to create a classification system for edge computing across three core areas:

  • Edge infrastructure
  • Edge devices
  • Edge software

The Taxonomy Project will draw upon the expertise of subject matter experts in each of the core areas to define and hopefully also visualize the complex relationships between the different components in edge computing. The Taxonomy Project WG is being led by Alex Marcham and the Open Glossary will publish the first taxonomies in Q3 2019.

The Edge Computing Landscape Map

At the request of The Linux Foundation, the State of the Edge project also contributed its Edge Computing Landscape Map, which is now a working group under the auspices of the Open Glossary. The Landscape WG is being led by Wesley Reisz and has just started out. Currently, they have regular weekly meetings (Tuesdays at noon Pacific Time) and the group is actively working on refining the categories the LF Edge Landscape. The group seeks wider participation and will be asking for help to test and validate the proposed categories.

You can see the landscape map evolve at https://landscape.lfedge.org and join the mailing list here.

Glossary 2.0

Edge computing is a rapidly evolving area of development, deployment and discussion, and the Open Glossary contributors aim to keep the glossary up to date with the latest developments in industry and academia, driven by contributions from our project members. During 2019, the Open Glossary team aims to release its 2.0 version of the Open Glossary, which will be a timely update that continues to form the basis for clear and concise communication on the edge.

To see open issues and pull requests or to make contributions, visit the GitHub repo. Go here to join the mailing list.

In Summary

2019 is shaping up to be a pivotal year for edge technologies, their proponents and of course their end users. The Open Glossary project, as part of The Linux Foundation, is dedicated to continuing its unique mission of bringing a single, open and definitive lingua franca to the world of edge computing. We hope you’ll join us during 2019 as we focus on these goals for the project, and look forward to your contributions.

Contributors can get involved with the Open Glossary project by joining the mailing lists and contributing and commenting on the GitHub repository for the project.

Alex Marcham is a technical marketing manager at Vapor IO. He is one of the primary contributors to the Open Glossary and leads the Taxonomy Project working group. Matt Trifiro is CMO of Vapor IO and is the Chair of the Open Glossary project.

Your Path to Edge Computing: Akraino Edge Stack’s Release 1

By | Akraino, Blog

By Kandan Kathirvel, Akraino Edge Stack TSC-Chair and Tina Tsou, Akraino Edge Stack TSC Co-Chair

The Akraino community was proud to announce the availability of its release 1 on June 6th. The community has experienced extremely rapid growth over the past year, in terms of both membership and community activity: Akraino includes broad contributions from across LF Edge, with 60% of LF Edge’s 60+ members contributing to project, as well as several other developers across the globe.

Before Akraino, developers had to download multiple open source software packages and integrate/test on deployable hardware, which prolonged innovation and increased cost. The Akraino community came up with a brilliant way to solve this integration challenge with the Blueprint model.

An Akraino Blueprint is not just a diagram; it’s real code that brings everything together so users can download and deploy the edge stack in their own environment to address a specific edge use case. Example use cases include IoT gateway, MEC for connected car, and a RAN intelligent controller that enables 5G infrastructure.

The Blueprints address interoperability, packaging, and testing under open standards, which reduces both overall deployment costs and integration time by users. The Akraino community will supply Blueprints across the LF Edge portfolio of projects, with plans to address 5G, IoT and a range of other edge use cases.

The key strength of the Akraino community is the well-defined process to welcome new Blueprints, new members, users and developers. The technical community is comprised of a Technical Steering committee (TSC), which consists of representatives from across member companies. The TSC acts as a “watchdog” to set process, monitor the community, and ensure open collaboration. In addition to the TSC, the Akraino community has seven sub-committees focused on much-needed areas such as security, edge APIs, CI and validation labs, upstream collaborations, documentation, process and community. Regular meetings are scheduled to ensure broader collaboration and accelerate progress on the various projects. The community calendar can be found here. It is not necessary to be a member to join the community calls, we invite anyone interested in learning more to join!

The above picture illustrates the primary use of Akraino R1 Blueprints and its targeted deployment areas. The release 1 Blueprints cover everything from a larger deployment in a telco-based edge cloud to a smaller deployment, such as in a public building like a stadium. Each Blueprint is validated via community standards on real physical lab hardware, hosted by either the community or the users.

Akraino Edge Stack prides itself on continuous refinement and development to ensure the success of Blueprints and projects. The community is already planning R2, which will include both new Blueprints and enhancements to existing Blueprints, tools for automated Blueprint validations, defined edge API’s, new community lab hardware, and much more. For future events and meetings please visit: https://wiki.akraino.org/display/AK/Akraino+TSC+Group+Calendar.

 

TelecomTV: Akraino Edge Stack makes its formal debut with telco-specific architecture blueprints for 5G and IoT

By | Akraino, In the News
  • Inaugural release unifies multiple sectors of the edge;
  • Release 1 delivers tested and validated deployment-ready blueprints;
  • Creating a framework for defining and standardising APIs across stacks;
  • Akraino is part of the LF Edge organisation

Forget the expanding universe, for telcos it is all about the ever-expanding network edge, which just got a little bit bigger yesterday with the news that the Akraino open source project has published its first software release. Although it was only launched in February 2018, Akraino had a rich pedigree with seed code from AT&T and has enjoyed plenty of support during its short time under the Linux Foundation’s LF Edge umbrella.

Akraino Edge Stack, to give the project its full name, is focused on creating an open source software stack that supports a high-availability cloud stack optimised for edge computing systems and applications. It is being designed to improve the state of edge cloud infrastructure for enterprise edge, OTT edge, as well as telecoms edge networks. The project promises to give users new levels of flexibility to scale their edge cloud services quickly, to maximise the applications and functions supported at the edge, and to help ensure the reliability of critical systems.

“Akraino Release 1 represents the first milestone towards creation of a much-needed common framework for edge solutions that address a diverse set of edge use cases,” said Arpit Joshipura, general manager, Networking, Automation, Edge and IoT, the Linux Foundation. “With the support of experts from all across the industry, we are paving the way to enable and support future technology at the edge.”

Read the full article here.

Hello System Management!

By | Blog, EdgeX Foundry

Written by Akram Ahmad, EdgeX Foundry contributor and Principal Software Engineer at Dell Technologies

For those of you not yet familiar with the canonical way of introducing new technology-centric stuff, at least the way we do it in the world of computer programming–and thinking here specifically to the “Hello World!” first-ever program introduced to the world by programming legends Kernighan and Ritchie with their C programming language–please allow me to clarify what may be an admittedly enigmatic title we’ve got for this blog post. Essentially, it was with the EdgeX Foundry Delhi Release that the team had the pleasure of introducing EdgeX System Management capability to the world! Hence, “Hello System Management!” (More on the Edinburgh Release in just a bit.)

It’s my ongoing privilege to be a part of helping design, implement, and shepherd System Management (or “SM” for short) to date, and going forward. With that in mind, I would like to give you a flavor of the capabilities that SM brings to the table.

You can think of the System Management Agent (SMA), in particular, as a brand-new service which serves as the coordinator for control plane information (i.e. status, configuration, and metrics for EdgeX services). The SMA also control actions on EdgeX services (i.e. starting, stopping, and restarting services). Cloud or third-party systems can, in turn, call on the API provided by the SMA to trigger the actions or to get the control plane data they need. In a nutshell, the SMA can serve as a one-stop shop for managing a deployed instance of EdgeX.

Each EdgeX micro service has a corresponding management API that the SMA calls on to help control that service (e.g. to stop the service) or fetch its latest configuration or metrics. The SMA, along with the management API provided by each service, will be expanded in future releases of EdgeX and will one day offer control plane data and actions via alternate protocols (for example via the well-known protocol SNMP that is part of the TCP/IP suite that powers the Internet as we know it today).

I invite you to hold on to the thought that, for the constellation of services that will be offered via EdgeX, there needs to be “controller” of sorts…

Now, let’s turn to the truism that an IoT platform like EdgeX is used to collect the data from “Things.” Put another way, the platform ingests data that is physically sensed from IoT sensors and devices. Work associated with collecting, managing, and disbursing sensed data is exactly the kind of work associated with a “data plane.” On the other hand, the kind of work associated with operating and managing the IoT platform software and infrastructure is best described as “control plane” operations.

This includes getting the IoT platform and infrastructure running (or shutdown), configuring the platform software for the particular use case, and understanding the health and status of the software platform (is it running and what type of resources is the IoT software platform using?). Analysis of any control plane data may be used to take action as well, but action revolves around the IoT platform itself–not the sensed or controlled world. For example, in the control plane, it may be determined that a service needs to be restarted because it is consuming too much memory.

This is where the SMA comes in!

The System Management (SM) service will assist in protecting EdgeX and reducing the surface area of an API attack. Rather than opening up access to all services to the central management system, the SM service serves as a single point proxy to the control plane for all of EdgeX services for the central management system. The SMA, therefore, reduces the number of access points to EdgeX and reduces potential security vulnerabilities. It also allows the central management system to be loosely coupled to all of EdgeX—requiring the central management system to again have just one access address (the address of the SM service) that it needs to know about for any EdgeX deployment.

Before digging deeper, let’s recap what we’ve learned so far: System Management (SM) functionality, as determined by the EdgeX community, is generally associated with control plane data and operations.  The control plane (and System Management) is about managing the IoT platform and infrastructure. The data plane is all about managing and understanding the physical world that the IoT platform is there to observe and control. Think about it: Whether one is talking about towering skyscrapers or flimsy tents rigged on the grounds of a park, there remains, as ever, the crucial need for control. Without coordination, things can get chaotic in a heartbeat.

Also, and crucially, SM is also about providing information—having retrieved that information in the first place—about the status of the services it manages. Eventually, building on this capability, SM will provide the means to reconfigure the services themselves. At this time, with the Edinburgh Release, SM can provide performance and memory metrics for requested services. Likewise, SM can provide detailed configuration information for the services requested by users of SM, as well as the health status of those services (whether given services are up or down.)

In other words, while control is a critical capability, SM is about more than just control. By the same token, we want to make it abundantly clear that we are building System Management (SM) capability to facilitate other central systems, and not be those central systems. In a nutshell, EdgeX SM is about helping promote interoperability—in this case, allowing you to manage EdgeX with your choice in central management system.

Let’s shift gears a bit now: When you look at a typical fog deployment, a larger management system will want to manage the control plane of the edge systems as well as all the intermediate and upper level nodes and resources of the overall deployment. Just as there is a management system to control all the nodes and infrastructure within a cloud data center, and across cloud data centers, so too there will likely be management systems that will manage and control all the nodes (from edge to cloud) and infrastructure of a complete fog or IoT deployment.

If you will be so kind as to allow me the use of just one more metaphor, it will be this one: Think to a team of workhorses ploughing the land (EdgeX services). Then think to the driver (System Management). Finally, and without going too crazy about the farming metaphor—all metaphors, including this one, can carry only so much water—I invite you to imagine two scenarios (1) First, the one without the other, and (2) Second, the two (i.e. the team of workhorses and the driver) working in unison. If you associated chaos with the first scenario, and clockwork unison with the second, you are in good company.

So with the Edinburgh Release, we will continue building SM capability to facilitate other central systems. Again, the goal is not to be those central systems, but rather to facilitate those systems. May your System Management (SM) learnings continue, and may the community be the better for it!

If you have questions or comments, visit the EdgeX Foundry Slack Channel and share your thoughts in the #community channel. Or, join the LF Edge Slack Channel and share your thoughts in the #EdgeX channel.

Project EVE Code Now Available

By | Blog, Project EVE

Project EVE (Edge Virtualization Engine), part of LF Edge since the organization’s inception, earlier this week marked an important milestone: the official handover of code from ZEDEDA. EVE provides an open standard for edge virtualization, helping make it as easy and secure to manage applications on edge devices as it is in the cloud. With EVE, enterprises can run a wide variety of applications on any edge-class gateway while enjoying the benefits of data center virtualization, like zero-touch provisioning and secure, one-click software update rollouts at IoT scale.

“Project EVE’s release under LF Edge is an important milestone for the edge computing industry,” said Melissa Evers-Hood, senior director of Google Operating Systems for Intel System Software Products, and chair of the LF Edge Governing Board. “An open approach to virtualization can help companies address the growth in diverse services and hardware configurations being deployed at the edge. Using virtualization to consolidate workloads provides companies with a more flexible and elastic infrastructure, allowing them to secure and manage these services while containing costs.”

Project EVE  allows applications ranging from legacy software programs running in virtual machines (VMs) to the latest microservices architectures to operate in a secure and reliable way on smaller edge devices. This is accomplished through the use of a type-1 hypervisor, an Edge Container runtime, and a hardened root-of-trust implementation, enabling workloads to run in either a VM or standard container environment. By decoupling software from hardware, EVE also allows for multi-tenant deployments that can operate in complete isolation from each other, increasing security and decreasing complexity.

Key features of Project EVE include:

  • Compatibility with all major edge hardware and cloud providers—no vendor lock-in
  • Ability to support any application that can run in a VM or standard container
  • Simplified application management through standardized APIs
  • Smarter hardware usage through coordinated resource allocation and partitioning
  • Ability to create a zero-trust approach to security, leveraging a hardened root-of-trust implementation

As the number of IoT devices continues to skyrocket, it’s becoming more and more important for businesses to be able to process, analyze, and act on sensor data in real time via local edge gateway systems. Project EVE provides a key component of the technology stack needed for powerful computing at the edge. By contributing the code for Project EVE to LF Edge, ZEDEDA is furthering the organization’s mission to create an open framework for edge computing.

For more information about Project EVE, visit https://www.lfedge.org/projects/eve/.

 

LF Edge Momentum Continues with Project EVE Seed Code, Project Demonstrations at IoT World and New Members

By | Akraino, Announcement, EdgeX Foundry, Project EVE

  • IoT OnPrem Edge Virtualization Engine seed code contributed by ZEDEDA to LF Edge
  • Four new members join existing community of 70+ LF Edge organizations
  • LF Edge on Display at IoT World, with Akraino Edge Stack, EdgeX Foundry and Project EVE demonstrations

SANTA CLARA, Calif. IoT World  – May 14, 2019 – LF Edge, an umbrella organization within the Linux Foundation that aims to establish an open, interoperable framework for edge computing independent of hardware, silicon, cloud, or operating system, today announced continued project momentum. Project Edge Virtualization Engine (EVE) receives initial seed code from LF Edge founding member ZEDEDA, as the community showcases a range of edge/IoT application demonstrations, from connected cars to wind turbines, on-site at IoT World.

Additionally, LF Edge welcomes new Associate and Liaison member organizations Industrial Internet Consortium (IIC), the LIONS Center at the Pennsylvania State University, OTAinfo, and University of New Hampshire’s Interoperability Lab (UNH-IOL).

“We are excited to see the LF community continue to collaborate on building unified edge solutions,” said Arpit Joshipura, general manager,  Networking, IoT and Edge Computing, the Linux Foundation. “We appreciate ZEDEDA’s leadership in helping us advance On-Prem Edge IoT with initiatives like Project EVE, and are eager to showcase the broad capabilities of LF Edge onsite in Santa Clara while welcoming our newest members.”

Edge Virtualization Engine Launches Initial Seed Code

Analysts predict more than 30B connected devices worldwide by 2020, all generating continuous data streams that will need to be processed quickly. Edge computing enables data processing through local gateway systems to reduce latency and provide faster response. With Project EVE, edge gateways and devices run a variety of edge workloads simultaneously, decoupling application management from the underlying hardware. Applications can be deployed in standard virtual machines (VM) or container environments and be managed through a standard set of APIs.

“With Project EVE, the goal is to create a single virtualization standard for edge devices for the industry to build around so that we can enjoy the benefits of cloud-native applications sooner rather than later,” said Said Ouissal, co-founder and CEO of ZEDEDA. “Imagine how much more impact we can achieve now that edge applications can be reliably managed and secured.”

LF Edge on Display at IoT World

LF Edge is on-site at IoT World in Santa Clara this week, May 13-16. Highlights include live project demonstrations in booth 610 and LF Edge community expert presentations:

  • Debut of the new Project EVE code
  • Live demonstrations of emerging Akraino blueprints for network cloud integration and industrial automation and Connected Cars
  • A smart build automation demo from EdgeX Foundry
  • A model wind turbine demonstrating EdgeX and Project EVE technology
  • Simulation of an industrial assembly line with sensor devices supporting MQTT and Monbus protocols
  • Arpit Joshipura will participate in two IoT panels; learn more: https://www.lfedge.org/event/iot-world-2019/.

More details on LF Edge’s activity at IoT World are available in this blog post. Please also follow @LF_Edge on twitter for real-time updates.

LF Edge Welcomes New Members 

LF Edge’s significant growth over the past five months– including strong industry event presence and face-to-face community meetings – continues today with the addition of new Associate and Liaison members:  Industrial Internet Consortium (IIC), the LIONS Center at the Pennsylvania State University, OTAinfo, and University of New Hampshire’s Interoperability Lab (UNH-IOL).

Comments from New Members:

“Through its liaison program, IIC works with industry liaison partners to build out the Industrial IoT ecosystem in an effort to accelerate IIoT adoption,” said Wael William Diab, Chair of the IIC Liaison Working Group and Secretary of the IIC Steering Committee. “Open source communities play an important role in the industry ecosystem. The expansion of the scope of liaison collaboration between IIC and LF Edge builds on the successful collaboration and further enables the digital transformation.”

“LF Edge stands as an indispensable bridge between industry and academia. It brings Open Source SD-WAN Frameworks to researchers,” said Peng Liu, Director of Lions Center at The Pennsylvania State University. “We, as security researchers, are eager to learn more, work more and make more contributions. We are currently working on verification of the designs, applicable NIDS solutions, end to end security validation of CI/CD of Akraino. With this precious partnership opportunity, we would like to contribute more to make these projects simple, secure, and scalable SD-WAN solutions.”

“Together, we have a chance to create history by developing the next gen security for connected devices,” said Nupur Mehta, Co-Founder and CEO of OTAinfo. “After revenue, security is the second biggest concern for all IoT companies, and yet there is no clear open source framework widely available. OTAinfo joined LF Edge to provide developers a comprehensive plug and play security framework that eliminates remote update vulnerabilities. We look forward to working with the Edge members and stakeholders in the connected community to revolutionize security on the Edge.”

“There is no denying the value proposition to virtualization and the move to open source partnerships within industry.  By bringing these topics together, and focusing on the network edge, the LF Edge projects are going to enable the next generation of really exciting applications.”  Said Lincoln Lavoie, Senior Engineer, Broadband Technologies at the University of New Hampshire InterOperability Laboratory “The UNH-IOL is thrilled to be participating in those projects, through our work with the Akraino project to develop the Community Lab, helping connect developers with the resources necessary to accelerate their open source efforts.”

About the Linux Foundation

Founded in 2000, the Linux Foundation is supported by more than 1,000 members and is the world’s leading home for collaboration on open source software, open standards, open data, and open hardware. Linux Foundation’s projects are critical to the world’s infrastructure including Linux, Kubernetes, Node.js and more. The Linux Foundation’s methodology focuses on leveraging best practices and addressing the needs of contributors, users and solution providers to create sustainable models for open collaboration. For more information, please visit us at linuxfoundation.org.

####

Forbes: Opening Up The Edge: Why Agnostic Cloud Software Is The Key To Edge Computing

By | In the News, Project EVE

The rising value of data has changed the business landscape forever, and companies have migrated en masse to the (theoretical) unlimited capacity of the cloud. As IoT devices have started to pump more data through networks and into the cloud, connected enterprises are managing their operations in data centers as well as using computational ability closer to the data source.

Read more at Forbes.