Skip to main content

Infinimesh IoT / IIoT platform is starting up!

Listen:

Today is a day we will never forget - infinimesh (https://www.infinimesh.io/) is starting and lifting off! Our Kubernetes, Apache Kafka ® and graph based Industrial IoT platform is entering the alpha stage! We have been working like maniacs over the past 14 months to bring a fully flexible, independent, patent and vendor lock-in free IoT platform to you! Soon it’s your chance to test and try it out, our closed alpha will be open for public on March 30, 2019 - Mark this date in your calendar!

An incredible platform comes to life

We believe smart and connected devices bring our society forward. Smart technology uses resources only when they are really necessary and thus prevents waste. On the other hand, when really required, smart things act and hence prevent accidents or simply enable a great user experience. We have started infinimesh 100% Open Source, without patents or closed software. Any software components we have developed, and to this we commit going forward, will be open - forever. Founded by engineers who built the backbone of the European Energy Revolution, infinimesh aims to make industrial and individual IoT secure, available and affordable for all. Infinimesh runs in all cloud offerings, be it public, hybrid or private. All you need is Linux; our platform works in any container environment as well as native.

Infinimesh on Google Cloud

We work with Google Cloud as strategic partner for our SaaS offering - and from today on the platform is running on GCP! Our SaaS offering, running in Google Cloud, is free for everybody up to 25 devices - ideally for makers, startups and industrial Proof of Concepts. That leaves enough room to bring ideas to live and test even larger installations and use the feature rich ecosystem of GCP to make your idea a successful product.

What can I do with infimesh IoT on GCP right now?

  • Connect devices securely via MQTT 3.1.1
  • Transfer desired and reported device states
  • Manage accounts (Create/Delete)
  • Manage Namespaces to organize devices and restrict access to devices
  • Create hierarchically organized objects, e.g. buildings, rooms to organize and model device hierarchies

How does it work?

Our Kubernetes Operator does the work a real operator would do: it not only installs the whole platform, but also takes care of required cloud/datacenter resources, updates, monitoring and handles incidents like errors. It attempts to resolve as many issues as possible on its own, and notifies human operators when human intervention is required. The operator is the glue between infinimesh and the target installation environment. Our alpha drop focuses on Google Cloud Platform and enables exactly this environment. More supported environments will follow.

Features

Device Management

Powerful but simple framework to visualize clusters of devices within your organization and set permissions up to device level.

Device Shadow

Real-time and two-way correspondence for every device in your fleet. Our highly-scalable backend can power millions of devices.

Timeseries Visualization

Great telemetry is based on timeseries. infinimesh has timeseries data capabilities built-in and enables meaningful monitoring.

Virtual Twins

A virtual twin is the digital copy of your physical asset. infinimesh provides virtual twins which give you the possibility to modify your physical device without even touching it

Machine Learning and AI

infinimesh has Machine Learning and Artificial Intelligence models built-in to rapidly detect anomalies and respond accordingly.

Roadmap and features ahead

OPC-UA with full open62541 support (binary protocol with encryption) and BACnet will be available within the next quarter.

OPC-UA is the leading semantic protocol for industry 4.0 and opens the full potential to industry proven stacks like Siemens MindSphere and IBM Watson for Industry. BACnet will also make its way into the platform quite soon, we expect a first drop in the next couple of weeks. BACnet is the most used communications protocol for Building Automation and Control (BAC) networks that leverage the ASHRAE, ANSI and ISO 16484-5 standard protocol and is used in various intelligent buildings as protocol stack.

What’s next?

More exciting news and announcements will follow in the next months, so use the platform and follow this blog or our channels to never miss news. We are happy to have you as user and customer and we will support you in any idea you have. Drop us a mail, or open a Feature Request (https://github.com/infinimesh/infinimesh/blob/master/.github/ISSUE_TEMPLATE/feature_request.md) or contact (https://infinimesh.io/contact.html#contact) us over our different channels - we are here.

Comments

Popular posts from this blog

Deal with corrupted messages in Apache Kafka

Under some strange circumstances it can happen that a message in a Kafka topic is corrupted. This happens often by using 3rd party frameworks together with Kafka. Additionally, Kafka < 0.9 has no lock at Log.read() at the consumer read level, but has a lock on Log.write(). This can cause a rare race condition, as described in KAKFA-2477 [1]. Probably a log entry looks like: ERROR Error processing message, stopping consumer: (kafka.tools.ConsoleConsumer$) kafka.message.InvalidMessageException: Message is corrupt (stored crc = xxxxxxxxxx, computed crc = yyyyyyyyyy Kafka-Tools Kafka stores the offset of every consumer in Zookeeper. To read out the offsets, Kafka provides handy tools [2]. But also zkCli.sh can be used, at least to display the consumer and the stored offsets. First we need to find the consumer for a topic (> Kafka 0.9): bin/kafka-consumer-groups.sh --zookeeper management01:2181 --describe --group test Prior to Kafka 0.9 the only possibility to get this inform

Hive query shows ERROR "too many counters"

A hive job face the odd " Too many counters:"  like Ended Job = job_xxxxxx with exception 'org.apache.hadoop.mapreduce.counters.LimitExceededException(Too many counters: 201 max=200)' FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MapRedTask Intercepting System.exit(1) These happens when operators are used in queries ( Hive Operators ). Hive creates 4 counters per operator, max upto 1000, plus a few additional counters like file read/write, partitions and tables. Hence the number of counter required is going to be dependent upon the query.  To avoid such exception, configure " mapreduce.job.counters.max " in mapreduce-site.xml to a value above 1000. Hive will fail when he is hitting the 1k counts, but other MR jobs not. A number around 1120 should be a good choice. Using " EXPLAIN EXTENDED " and " grep -ri operators | wc -l " print out the used numbers of operators. Use this value to tweak the MR s

GPT & GenAI for Startup Storytelling

OpenAI and Bard   are the most used GenAI tools today; the first one has a massive Microsoft investment, and the other one is an experiment from Google. But did you know that you can also use them to optimize and hack your startup?  For startups, creating pitch scripts, sales emails, and elevator pitches with generative AI (GenAI) can help you not only save time but also validate your marketing and wording. Curious? Here are a few prompt hacks for startups to create,improve, and validate buyer personas, your startup's mission/vision statements, and unique selling proposition (USP) definitions. First Step: Introduce yourself and your startup Introduce yourself, your startup, your website, your idea, your position, and in a few words what you are doing to the chatbot: Prompt : I'm NAME and our startup NAME, with website URL, is doing WHATEVER. With PRODUCT NAME, we aim to change or disrupt INDUSTRY. Bard is able to pull information from your website. I'm not sure if ChatGPT