Publications
-
Rethinking Product Management: Flexibility and Customer Obsession for Success
Novatechflow
The right product development and management methodology always depends on your goals and geography. But if you want to make insane growth happen, obsess over your customers, find the right balance between strategy and flexibility, and implement a culture of constant learning.
-
AI & ML for Business Efficiency
Medium
Strategic AI implementation enables hyper-personalization and proactive service to exceed customer expectations. It also attracts top talent interested in working on cutting-edge projects and augments human capabilities.
Artificial intelligence (AI) and machine learning (ML) have become inescapable tech buzzwords. Yet, for many businesses, the actual value of these technologies can still feel out of reach. Is the cost of inventory mishaps eating away at your profit margins? Or are you…Strategic AI implementation enables hyper-personalization and proactive service to exceed customer expectations. It also attracts top talent interested in working on cutting-edge projects and augments human capabilities.
Artificial intelligence (AI) and machine learning (ML) have become inescapable tech buzzwords. Yet, for many businesses, the actual value of these technologies can still feel out of reach. Is the cost of inventory mishaps eating away at your profit margins? Or are you struggling to keep up with changing customer expectations? It's time to move past the hype and focus on how AI/ML can deliver tangible improvements and drive ROI in the real world. -
ETL vs. ELT: Decoding the Data Wrangling Showdown for Your Next Project
Scalytics
Transforming raw data into actionable insights requires the right approach. ETL (Extract, Transform, Load) and ELT (Extract, Load, Transform) are the leading methods for streamlining data preparation. ETL ensures upfront data quality, ideal for regulated industries or structured data. ELT prioritizes speed and flexibility for rapid insights. This blog post offers a clear breakdown of ETL and ELT, guiding developers on choosing the right approach based on project needs, data types, and…
Transforming raw data into actionable insights requires the right approach. ETL (Extract, Transform, Load) and ELT (Extract, Load, Transform) are the leading methods for streamlining data preparation. ETL ensures upfront data quality, ideal for regulated industries or structured data. ELT prioritizes speed and flexibility for rapid insights. This blog post offers a clear breakdown of ETL and ELT, guiding developers on choosing the right approach based on project needs, data types, and compliance. We also explore the future of data integration, highlighting the potential of hybrid models, real-time transformation, and federated learning for secure AI development.
-
How Mature is Your Data? Assess Your Business' Data Health Now
Scalytics
Understand data maturity's importance for modern businesses. Discover why your business might lag and how to advance with data handling and analytics. This post explores the concept of data maturity and its crucial role in today's data-driven business landscape. We'll delve into the key areas businesses need to focus on, the reasons some fall behind, and how to bridge the gap.
-
Modern Product Management
Medium
Modern product management isn’t just about adding features or functionalities. It’s a holistic approach that puts the digital customer at the center of everything you do. This means understanding your customers and their ever-evolving needs. To do this, you’ll need to dive deep into your customers’ behavior, preferences and pain points through multiple channels, such as customer research, data analysis and feedback mechanisms.
By building a culture of empathy with your customers, your…Modern product management isn’t just about adding features or functionalities. It’s a holistic approach that puts the digital customer at the center of everything you do. This means understanding your customers and their ever-evolving needs. To do this, you’ll need to dive deep into your customers’ behavior, preferences and pain points through multiple channels, such as customer research, data analysis and feedback mechanisms.
By building a culture of empathy with your customers, your product managers will be able to translate your insights into actionable solutions that will drive product development and improve your overall customer experience (CX). -
Data Silos are Killing Your AI Performance
DataBloom Blog
Organizations investing in analytics, artificial intelligence (AI), and other data-driven efforts face a rising challenge: a lack of integration across data sources, which limits their ability to extract actual value from these investments. To enable greater business insights, IT and business leaders must eliminate these data silos, some of which are operational and others of which are cultural. A large percentage of organizations and their leadership teams understand the value of data and are…
Organizations investing in analytics, artificial intelligence (AI), and other data-driven efforts face a rising challenge: a lack of integration across data sources, which limits their ability to extract actual value from these investments. To enable greater business insights, IT and business leaders must eliminate these data silos, some of which are operational and others of which are cultural. A large percentage of organizations and their leadership teams understand the value of data and are working to develop a modern data strategy.
-
How Databloom.ai and the Blossom Development Environment Are Revolutionizing Data Science
Medium
Databloom.ai and the Blossom Development Environment are revolutionizing the field of data science by providing powerful tools and workflows to streamline data analysis and modeling. These tools enable data scientists to work more efficiently and effectively, leading to faster insights and better decision-making. Learn more about how Databloom.ai and the Blossom Development Environment are changing the game in our latest article.
-
Federated Learning is necessary to avoid bias in Generative AI
tech.mt
Generative AI is a rapidly advancing field that holds the promise of revolutionizing the way we interact with technology. From generating high-quality digital images to creating realistic videos, or NLP-based text and information processing algorithms, the potential applications are endless. However, as we all know, with any new technology comes ethical concerns and the obligation to ensure that it is used for the greater good. One, if not the most threatening, of the significant challenges…
Generative AI is a rapidly advancing field that holds the promise of revolutionizing the way we interact with technology. From generating high-quality digital images to creating realistic videos, or NLP-based text and information processing algorithms, the potential applications are endless. However, as we all know, with any new technology comes ethical concerns and the obligation to ensure that it is used for the greater good. One, if not the most threatening, of the significant challenges posed by generative AI is the risk of bias in the algorithms and models that it creates.
-
Predicting and modeling events using generative AI, federated learning, and digital twins
databloom.ai
Create more accurate digital twin models, which can lead to better predictions and more efficient maintenance of assets. Digital twin models can also be trained using federated learning, by combining data from different sensors and systems.
-
How can Federated Learning (FL) boost your company's digital transformation?
databloom.ai
Federated Learning (FL) is a machine learning technique where a centralized model is trained on decentralized data, that is, data that is distributed across multiple devices, such as smartphones, laptops, and IoT devices.
-
Are You Wasting Money in the Cloud?
tech.mt
Public cloud displacement is a subject we don’t discuss as often as we ought to. Many see relocating data and apps back from a public cloud provider to an enterprise data center as an admission that the initial decision to move the workloads to the cloud was a grave error. In my opinion, this is less of a failure than a hosting platform change depending on the current state of the economy. People frequently return to more conventional platforms because of the high expense of cloud computing…
Public cloud displacement is a subject we don’t discuss as often as we ought to. Many see relocating data and apps back from a public cloud provider to an enterprise data center as an admission that the initial decision to move the workloads to the cloud was a grave error. In my opinion, this is less of a failure than a hosting platform change depending on the current state of the economy. People frequently return to more conventional platforms because of the high expense of cloud computing. You surly remember the article from Dropbox, explaining why they left their public infrastructure and went to a private cloud approach.
-
What exactly is Federated Learning, and why is it so crucial?
databloom.ai
Federated Learning (FL) is a machine learning method in which a model is trained on multiple devices, such as smartphones or edge devices, rather than a centralized server. The devices, also known as clients, train a model on their own data and then send updates to a central server. After that, the server aggregates the updates and returns the improved model to the clients. This procedure is repeated until the model achieves the desired level of precision.
-
The Data Mesh - should you adapt?
databloom.ai
As it focuses on delivering useful and safe data products, Data Mesh is a strategic approach to modern data management and a strategy to support an organization's journey toward digital transformation. Data Mesh's major goal is to advance beyond the established centralized data management techniques of using data warehouses and data lakes. By giving data producers and data consumers the ability to access and handle data without having to go through the hassle of involving the data lake or data…
As it focuses on delivering useful and safe data products, Data Mesh is a strategic approach to modern data management and a strategy to support an organization's journey toward digital transformation. Data Mesh's major goal is to advance beyond the established centralized data management techniques of using data warehouses and data lakes. By giving data producers and data consumers the ability to access and handle data without having to go through the hassle of involving the data lake or data warehouse team, Data Mesh highlights the concept of organizational agility. Data Mesh's decentralized approach distributes data ownership to industry-specific organizations that use, control, and manage data as a product.
-
Combined Federated Data Services with Blossom and Flower
databloom.ai
How to build a chatbot system, which serves multiple functions and customers across the world, like in a bank? A chatbot stack typically uses NLP combined with multiple data source to provide a natural communication between humans and machines. The demand of Machine-Human interaction and human based communication has considerably increased and the forecasts of Gartner are a testament to it.
-
Scalable Timeseries with Scalytics, Redis and Grafana
Medium
Scalytics combines our two open source projects wayang and infinimesh into one AI-IoT solution, fully open source and ultra-scalable. Our AIoT platform is built to make data privacy and data ownership possible, without any doubt. To achieve this, we have built our platform in a cloud-native way and entirely API driven. It allows our customers to integrate our cloud into their systems without compromising IT Security or even move to a public cloud provider. We understand ourselves as a stretched…
Scalytics combines our two open source projects wayang and infinimesh into one AI-IoT solution, fully open source and ultra-scalable. Our AIoT platform is built to make data privacy and data ownership possible, without any doubt. To achieve this, we have built our platform in a cloud-native way and entirely API driven. It allows our customers to integrate our cloud into their systems without compromising IT Security or even move to a public cloud provider. We understand ourselves as a stretched workbench for any IoT related ideas our customers might have.
-
Enabling IoT to Establish a Sustainable Value Chain
CIO Applications Europe
IoT devices are getting more and more intelligent and can now create meshed networks by itself, switching from a sensor into an actor and transferring informations only for the meshed neighbors. For example a connected car could tell the future home that the homeowner will be at home in 5 minutes and the garage door and the door need to be unlocked in time, the lights need to be switched on and the grid operator needs to be informed that the wallbox now charges with 22KW. In near future this…
IoT devices are getting more and more intelligent and can now create meshed networks by itself, switching from a sensor into an actor and transferring informations only for the meshed neighbors. For example a connected car could tell the future home that the homeowner will be at home in 5 minutes and the garage door and the door need to be unlocked in time, the lights need to be switched on and the grid operator needs to be informed that the wallbox now charges with 22KW. In near future this will happen over direct meshed information cells, operated by always connected devices, wearables, sensors, actors, mobile devices - short: everything. And all cloud provider offer dozens of solution to master the challenges, on the one, other or complete different way.
-
Big data need not mean high costs and lengthy training cycles
Techerati
These days it means more to understand business processes and to transform them into data-driven opportunities – using cloud technology when needed
For a lot, but certainly not all enterprises, a typical big data software is Apache Hadoop. Hadoop is the new legacy standard in enterprises when it comes to on-premise tools. Cloud-based big data is still fairly new for enterprises and cloud providers, as the technology has only risen up to become cloud-ready in the last two years or so…These days it means more to understand business processes and to transform them into data-driven opportunities – using cloud technology when needed
For a lot, but certainly not all enterprises, a typical big data software is Apache Hadoop. Hadoop is the new legacy standard in enterprises when it comes to on-premise tools. Cloud-based big data is still fairly new for enterprises and cloud providers, as the technology has only risen up to become cloud-ready in the last two years or so. Popular tools here are Apache Kafka, Apache Spark or Apache Flink – these are all run best on bare metal servers. -
The next stage of BigData
Blog
Right now, the terms BigData and Hadoop are used as one and the same - often like the buzzword of buzzwords. And they sound mostly as a last time call, often made by agencies to convince people to start the Hadoop journey before the train leaves the station. Don’t fall into that trap.
-
Shifting paradigms in the world of BigData
Blog
In building the next generation of applications, companies and stakeholders need to adopt new paradigms. The need for this shift is predicated on the fundamental belief that building a new application at scale requires tailored solutions to that application’s unique challenges, business model and ROI.
-
Open Source based Hyper-Converged Infrastructures and Hadoop
Blog
Hyper-Converged Infrastructures are used by more than 50% of the interviewed businesses, tendentious increasing. But what does this mean for BigData solutions, and Hadoop especially? What tools and technologies can be used, what are the limitations and the gains from such a solution?
To build a production ready and reliable private cloud to support Hadoop clusters as well as on-demand and static I have made great experience with OpenStack, Saltstack and the Sahara plugin for Openstack. -
SolR, NiFi, Twitter and CDH 5.7
Blog
This demo shows that's pretty easy today by using available tools to setup more or less complex data flows within a few hours. Apache NiFi is pretty stable, has a lot of sinks available and runs now 2 weeks in Google Compute, captured over 200 mio tweets and stored them in SolR as well as in HDFS.
-
The Ultimate Healthcare Big Data Dictionary
Evariant Ask Eva
To shed light on big data and its implications in the healthcare industry, we have compiled a list of big data related terms, along with their definitions. We hope it proves helpful in better understanding the scope of healthcare big data, as well as how to turn big data into practical data and insights that can lead to more successful campaigns.
-
Hadoop based SQL engines
Blog
Apache Hadoop comes more and more into the focus of business critical architectures and applications. Naturally SQL based solutions are the first to get considered, but the market is evolving and new tools are coming up, but leaving unnoticed.
-
Hadoop server performance tuning
Blog
To tune a Hadoop cluster from a DevOps perspective needs an understanding of the kernel principles and linux. The following article will describe the most important parameters together with tricks for an optimal tuning.
-
Facebook's Presto
Blog
In November 2013 Facebook published their Presto engine as Open Source, available at GitHub. Presto is a distributed interactive SQL query engine, able to run over dozens of modern BigData stores, based on Apache Hive or Cassandra. Presto comes with a limited JDBC Connector, supports Hive 0.13 with Parquet and Views.
-
How HDFS protects your data
Blog
HDFS is designed to protect data in different ways to minimize the risk of data loss with a valuable write speed. This enables in some circumstances HDFS as a NAS replacement for large files with the possibility to quickly access the stored data.
-
Impala and Kerberos
Blog
Impala provides fast, interactive SQL queries directly on your Apache Hadoop data stored in HDFS or HBase. In addition to using the same unified storage platform, Impala also uses the same metadata, SQL syntax (Hive SQL), ODBC driver and user interface (Hue Beeswax) as Apache Hive.
-
BigData - Eine Uebersicht
Blog
Overview about the Apache Hadoop Ecosystem, written in german
-
FlumeNG
Blog
Flume, the decentralized log collector, makes some great progress. Since the project has reached an Apache incubating tier the development on the next generation (NG) has reached a significant level.
-
Use snappy codec in Hive
Blog
Snappy is a compression and decompression library, initially developed from Google and now integrated into Hadoop. Snappy acts about 10% faster than LZO, the biggest differences are the packaging and that snappy only provides a codec and does not have a container spec, whereas LZO has a file-format container and a compression codec.
-
Export HDFS over CIFS
Blog
For some reasons it could be a good idea to make a hdfs filesystem available across networks as a exported share. Here I describe a working scenario with linux and hadoop with tools both have on board.
Projects
-
Apache Wayang
Apache Wayang is a system designed to fully support cross-platform data processing: It enables users to run data analytics over multiple data processing platforms. For this, it provides an abstraction on top of existing platforms in order to run data analytic tasks on top of any set of platforms. As a result, users can focus on the logics of their applications rather on the intricacies of the underlying platforms.
-
infinimesh
infinimesh is a opinionated multi-tenant hyperscale Internet of Things platform to connect IoT devices fast and securely with minimal TCO. It features a unique Graph-based authorization system, allowing users & engineers to create arbitrary hierarchical ontologies, with the possibility to scope permissions down to single sub-devices to specific users (e.g. suppliers). It exposes simple to consume RESTful & gRPC APIs with both high-level (e.g. device shadow) and low-level (sending messages)…
infinimesh is a opinionated multi-tenant hyperscale Internet of Things platform to connect IoT devices fast and securely with minimal TCO. It features a unique Graph-based authorization system, allowing users & engineers to create arbitrary hierarchical ontologies, with the possibility to scope permissions down to single sub-devices to specific users (e.g. suppliers). It exposes simple to consume RESTful & gRPC APIs with both high-level (e.g. device shadow) and low-level (sending messages) concepts. The infinimesh IoT platform is open source and fully kubernetes compliant. No vendor lock-in - run it yourself on Kubernetes in your own datacenter, under your control with maximum data privacy.
-
Apache Hadoop
-
The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage. Rather than rely on hardware to deliver high-availability, the library itself is designed to detect and handle failures at the application layer, so delivering a highly-available service on top of a cluster…
The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage. Rather than rely on hardware to deliver high-availability, the library itself is designed to detect and handle failures at the application layer, so delivering a highly-available service on top of a cluster of computers, each of which may be prone to failures.
-
Apache Flume
-
Flume is a distributed, reliable, and available service for efficiently collecting, aggregating, and moving large amounts of log data. It has a simple and flexible architecture based on streaming data flows. It is robust and fault tolerant with tunable reliability mechanisms and many failover and recovery mechanisms. It uses a simple extensible data model that allows for online analytic application.
Explore collaborative articles
We’re unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.
Explore MoreOthers named Alexander Alten
-
Alexander Alten
immer lösungsorientiert
-
Alexander Alten
-
Alex Alten
-
Dr. med. Tim Alexander Alten
Sektionsleiter Radiologie bei St. Vincenz Kliniken
7 others named Alexander Alten are on LinkedIn
See others named Alexander Alten