Alexander Alten

Mellieha, Malta Contact Info
3K followers 500+ connections

Join to view profile

Publications

  • Rethinking Product Management: Flexibility and Customer Obsession for Success

    Novatechflow

    The right product development and management methodology always depends on your goals and geography. But if you want to make insane growth happen, obsess over your customers, find the right balance between strategy and flexibility, and implement a culture of constant learning.

    See publication
  • AI & ML for Business Efficiency

    Medium

    Strategic AI implementation enables hyper-personalization and proactive service to exceed customer expectations. It also attracts top talent interested in working on cutting-edge projects and augments human capabilities.

    Artificial intelligence (AI) and machine learning (ML) have become inescapable tech buzzwords. Yet, for many businesses, the actual value of these technologies can still feel out of reach. Is the cost of inventory mishaps eating away at your profit margins? Or are you…

    Strategic AI implementation enables hyper-personalization and proactive service to exceed customer expectations. It also attracts top talent interested in working on cutting-edge projects and augments human capabilities.

    Artificial intelligence (AI) and machine learning (ML) have become inescapable tech buzzwords. Yet, for many businesses, the actual value of these technologies can still feel out of reach. Is the cost of inventory mishaps eating away at your profit margins? Or are you struggling to keep up with changing customer expectations? It's time to move past the hype and focus on how AI/ML can deliver tangible improvements and drive ROI in the real world.

    See publication
  • ETL vs. ELT: Decoding the Data Wrangling Showdown for Your Next Project

    Scalytics

    Transforming raw data into actionable insights requires the right approach. ETL (Extract, Transform, Load) and ELT (Extract, Load, Transform) are the leading methods for streamlining data preparation. ETL ensures upfront data quality, ideal for regulated industries or structured data. ELT prioritizes speed and flexibility for rapid insights. This blog post offers a clear breakdown of ETL and ELT, guiding developers on choosing the right approach based on project needs, data types, and…

    Transforming raw data into actionable insights requires the right approach. ETL (Extract, Transform, Load) and ELT (Extract, Load, Transform) are the leading methods for streamlining data preparation. ETL ensures upfront data quality, ideal for regulated industries or structured data. ELT prioritizes speed and flexibility for rapid insights. This blog post offers a clear breakdown of ETL and ELT, guiding developers on choosing the right approach based on project needs, data types, and compliance. We also explore the future of data integration, highlighting the potential of hybrid models, real-time transformation, and federated learning for secure AI development.

    See publication
  • How Mature is Your Data? Assess Your Business' Data Health Now

    Scalytics

    Understand data maturity's importance for modern businesses. Discover why your business might lag and how to advance with data handling and analytics. This post explores the concept of data maturity and its crucial role in today's data-driven business landscape. We'll delve into the key areas businesses need to focus on, the reasons some fall behind, and how to bridge the gap.

    See publication
  • Modern Product Management

    Medium

    Modern product management isn’t just about adding features or functionalities. It’s a holistic approach that puts the digital customer at the center of everything you do. This means understanding your customers and their ever-evolving needs. To do this, you’ll need to dive deep into your customers’ behavior, preferences and pain points through multiple channels, such as customer research, data analysis and feedback mechanisms.
    By building a culture of empathy with your customers, your…

    Modern product management isn’t just about adding features or functionalities. It’s a holistic approach that puts the digital customer at the center of everything you do. This means understanding your customers and their ever-evolving needs. To do this, you’ll need to dive deep into your customers’ behavior, preferences and pain points through multiple channels, such as customer research, data analysis and feedback mechanisms.
    By building a culture of empathy with your customers, your product managers will be able to translate your insights into actionable solutions that will drive product development and improve your overall customer experience (CX).

    See publication
  • Data Silos are Killing Your AI Performance

    DataBloom Blog

    Organizations investing in analytics, artificial intelligence (AI), and other data-driven efforts face a rising challenge: a lack of integration across data sources, which limits their ability to extract actual value from these investments. To enable greater business insights, IT and business leaders must eliminate these data silos, some of which are operational and others of which are cultural. A large percentage of organizations and their leadership teams understand the value of data and are…

    Organizations investing in analytics, artificial intelligence (AI), and other data-driven efforts face a rising challenge: a lack of integration across data sources, which limits their ability to extract actual value from these investments. To enable greater business insights, IT and business leaders must eliminate these data silos, some of which are operational and others of which are cultural. A large percentage of organizations and their leadership teams understand the value of data and are working to develop a modern data strategy.

    See publication
  • How Databloom.ai and the Blossom Development Environment Are Revolutionizing Data Science

    Medium

    Databloom.ai and the Blossom Development Environment are revolutionizing the field of data science by providing powerful tools and workflows to streamline data analysis and modeling. These tools enable data scientists to work more efficiently and effectively, leading to faster insights and better decision-making. Learn more about how Databloom.ai and the Blossom Development Environment are changing the game in our latest article.

    See publication
  • Federated Learning is necessary to avoid bias in Generative AI

    tech.mt

    Generative AI is a rapidly advancing field that holds the promise of revolutionizing the way we interact with technology. From generating high-quality digital images to creating realistic videos, or NLP-based text and information processing algorithms, the potential applications are endless. However, as we all know, with any new technology comes ethical concerns and the obligation to ensure that it is used for the greater good. One, if not the most threatening, of the significant challenges…

    Generative AI is a rapidly advancing field that holds the promise of revolutionizing the way we interact with technology. From generating high-quality digital images to creating realistic videos, or NLP-based text and information processing algorithms, the potential applications are endless. However, as we all know, with any new technology comes ethical concerns and the obligation to ensure that it is used for the greater good. One, if not the most threatening, of the significant challenges posed by generative AI is the risk of bias in the algorithms and models that it creates.‍

    See publication
  • Predicting and modeling events using generative AI, federated learning, and digital twins

    databloom.ai

    Create more accurate digital twin models, which can lead to better predictions and more efficient maintenance of assets. Digital twin models can also be trained using federated learning, by combining data from different sensors and systems.

    See publication
  • How can Federated Learning (FL) boost your company's digital transformation?

    databloom.ai

    Federated Learning (FL) is a machine learning technique where a centralized model is trained on decentralized data, that is, data that is distributed across multiple devices, such as smartphones, laptops, and IoT devices.

    See publication
  • Are You Wasting Money in the Cloud?

    tech.mt

    Public cloud displacement is a subject we don’t discuss as often as we ought to. Many see relocating data and apps back from a public cloud provider to an enterprise data center as an admission that the initial decision to move the workloads to the cloud was a grave error. In my opinion, this is less of a failure than a hosting platform change depending on the current state of the economy. People frequently return to more conventional platforms because of the high expense of cloud computing…

    Public cloud displacement is a subject we don’t discuss as often as we ought to. Many see relocating data and apps back from a public cloud provider to an enterprise data center as an admission that the initial decision to move the workloads to the cloud was a grave error. In my opinion, this is less of a failure than a hosting platform change depending on the current state of the economy. People frequently return to more conventional platforms because of the high expense of cloud computing. You surly remember the article from Dropbox, explaining why they left their public infrastructure and went to a private cloud approach.

    See publication
  • What exactly is Federated Learning, and why is it so crucial?

    databloom.ai

    Federated Learning (FL) is a machine learning method in which a model is trained on multiple devices, such as smartphones or edge devices, rather than a centralized server. The devices, also known as clients, train a model on their own data and then send updates to a central server. After that, the server aggregates the updates and returns the improved model to the clients. This procedure is repeated until the model achieves the desired level of precision.

    See publication
  • The Data Mesh - should you adapt?

    databloom.ai

    As it focuses on delivering useful and safe data products, Data Mesh is a strategic approach to modern data management and a strategy to support an organization's journey toward digital transformation. Data Mesh's major goal is to advance beyond the established centralized data management techniques of using data warehouses and data lakes. By giving data producers and data consumers the ability to access and handle data without having to go through the hassle of involving the data lake or data…

    As it focuses on delivering useful and safe data products, Data Mesh is a strategic approach to modern data management and a strategy to support an organization's journey toward digital transformation. Data Mesh's major goal is to advance beyond the established centralized data management techniques of using data warehouses and data lakes. By giving data producers and data consumers the ability to access and handle data without having to go through the hassle of involving the data lake or data warehouse team, Data Mesh highlights the concept of organizational agility. Data Mesh's decentralized approach distributes data ownership to industry-specific organizations that use, control, and manage data as a product.

    See publication
  • Combined Federated Data Services with Blossom and Flower

    databloom.ai

    How to build a chatbot system, which serves multiple functions and customers across the world, like in a bank? A chatbot stack typically uses NLP combined with multiple data source to provide a natural communication between humans and machines. The demand of Machine-Human interaction and human based communication has considerably increased and the forecasts of Gartner are a testament to it.

    See publication
  • Scalable Timeseries with Scalytics, Redis and Grafana

    Medium

    Scalytics combines our two open source projects wayang and infinimesh into one AI-IoT solution, fully open source and ultra-scalable. Our AIoT platform is built to make data privacy and data ownership possible, without any doubt. To achieve this, we have built our platform in a cloud-native way and entirely API driven. It allows our customers to integrate our cloud into their systems without compromising IT Security or even move to a public cloud provider. We understand ourselves as a stretched…

    Scalytics combines our two open source projects wayang and infinimesh into one AI-IoT solution, fully open source and ultra-scalable. Our AIoT platform is built to make data privacy and data ownership possible, without any doubt. To achieve this, we have built our platform in a cloud-native way and entirely API driven. It allows our customers to integrate our cloud into their systems without compromising IT Security or even move to a public cloud provider. We understand ourselves as a stretched workbench for any IoT related ideas our customers might have.

    See publication
  • Enabling IoT to Establish a Sustainable Value Chain

    CIO Applications Europe

    IoT devices are getting more and more intelligent and can now create meshed networks by itself, switching from a sensor into an actor and transferring informations only for the meshed neighbors. For example a connected car could tell the future home that the homeowner will be at home in 5 minutes and the garage door and the door need to be unlocked in time, the lights need to be switched on and the grid operator needs to be informed that the wallbox now charges with 22KW. In near future this…

    IoT devices are getting more and more intelligent and can now create meshed networks by itself, switching from a sensor into an actor and transferring informations only for the meshed neighbors. For example a connected car could tell the future home that the homeowner will be at home in 5 minutes and the garage door and the door need to be unlocked in time, the lights need to be switched on and the grid operator needs to be informed that the wallbox now charges with 22KW. In near future this will happen over direct meshed information cells, operated by always connected devices, wearables, sensors, actors, mobile devices - short: everything. And all cloud provider offer dozens of solution to master the challenges, on the one, other or complete different way.

    See publication
  • Big data need not mean high costs and lengthy training cycles

    Techerati

    These days it means more to understand business processes and to transform them into data-driven opportunities – using cloud technology when needed

    For a lot, but certainly not all enterprises, a typical big data software is Apache Hadoop. Hadoop is the new legacy standard in enterprises when it comes to on-premise tools. Cloud-based big data is still fairly new for enterprises and cloud providers, as the technology has only risen up to become cloud-ready in the last two years or so…

    These days it means more to understand business processes and to transform them into data-driven opportunities – using cloud technology when needed

    For a lot, but certainly not all enterprises, a typical big data software is Apache Hadoop. Hadoop is the new legacy standard in enterprises when it comes to on-premise tools. Cloud-based big data is still fairly new for enterprises and cloud providers, as the technology has only risen up to become cloud-ready in the last two years or so. Popular tools here are Apache Kafka, Apache Spark or Apache Flink – these are all run best on bare metal servers.

    See publication
  • The next stage of BigData

    Blog

    Right now, the terms BigData and Hadoop are used as one and the same - often like the buzzword of buzzwords. And they sound mostly as a last time call, often made by agencies to convince people to start the Hadoop journey before the train leaves the station. Don’t fall into that trap.

    See publication
  • Shifting paradigms in the world of BigData

    Blog

    In building the next generation of applications, companies and stakeholders need to adopt new paradigms. The need for this shift is predicated on the fundamental belief that building a new application at scale requires tailored solutions to that application’s unique challenges, business model and ROI.

    See publication
  • Open Source based Hyper-Converged Infrastructures and Hadoop

    Blog

    Hyper-Converged Infrastructures are used by more than 50% of the interviewed businesses, tendentious increasing. But what does this mean for BigData solutions, and Hadoop especially? What tools and technologies can be used, what are the limitations and the gains from such a solution?

    To build a production ready and reliable private cloud to support Hadoop clusters as well as on-demand and static I have made great experience with OpenStack, Saltstack and the Sahara plugin for Openstack.

    See publication
  • SolR, NiFi, Twitter and CDH 5.7

    Blog

    This demo shows that's pretty easy today by using available tools to setup more or less complex data flows within a few hours. Apache NiFi is pretty stable, has a lot of sinks available and runs now 2 weeks in Google Compute, captured over 200 mio tweets and stored them in SolR as well as in HDFS.

    See publication
  • The Ultimate Healthcare Big Data Dictionary

    Evariant Ask Eva

    To shed light on big data and its implications in the healthcare industry, we have compiled a list of big data related terms, along with their definitions. We hope it proves helpful in better understanding the scope of healthcare big data, as well as how to turn big data into practical data and insights that can lead to more successful campaigns.

    See publication
  • Hadoop based SQL engines

    Blog

    Apache Hadoop comes more and more into the focus of business critical architectures and applications. Naturally SQL based solutions are the first to get considered, but the market is evolving and new tools are coming up, but leaving unnoticed.

    See publication
  • Hadoop server performance tuning

    Blog

    To tune a Hadoop cluster from a DevOps perspective needs an understanding of the kernel principles and linux. The following article will describe the most important parameters together with tricks for an optimal tuning.

    See publication
  • Facebook's Presto

    Blog

    In November 2013 Facebook published their Presto engine as Open Source, available at GitHub. Presto is a distributed interactive SQL query engine, able to run over dozens of modern BigData stores, based on Apache Hive or Cassandra. Presto comes with a limited JDBC Connector, supports Hive 0.13 with Parquet and Views.

    See publication
  • How HDFS protects your data

    Blog

    HDFS is designed to protect data in different ways to minimize the risk of data loss with a valuable write speed. This enables in some circumstances HDFS as a NAS replacement for large files with the possibility to quickly access the stored data.

    See publication
  • Impala and Kerberos

    Blog

    Impala provides fast, interactive SQL queries directly on your Apache Hadoop data stored in HDFS or HBase. In addition to using the same unified storage platform, Impala also uses the same metadata, SQL syntax (Hive SQL), ODBC driver and user interface (Hue Beeswax) as Apache Hive.

    See publication
  • BigData - Eine Uebersicht

    Blog

    Overview about the Apache Hadoop Ecosystem, written in german

    See publication
  • FlumeNG

    Blog

    Flume, the decentralized log collector, makes some great progress. Since the project has reached an Apache incubating tier the development on the next generation (NG) has reached a significant level.

    See publication
  • Use snappy codec in Hive

    Blog

    Snappy is a compression and decompression library, initially developed from Google and now integrated into Hadoop. Snappy acts about 10% faster than LZO, the biggest differences are the packaging and that snappy only provides a codec and does not have a container spec, whereas LZO has a file-format container and a compression codec.

    See publication
  • Export HDFS over CIFS

    Blog

    For some reasons it could be a good idea to make a hdfs filesystem available across networks as a exported share. Here I describe a working scenario with linux and hadoop with tools both have on board.

    See publication

Projects

  • Apache Wayang

    Apache Wayang is a system designed to fully support cross-platform data processing: It enables users to run data analytics over multiple data processing platforms. For this, it provides an abstraction on top of existing platforms in order to run data analytic tasks on top of any set of platforms. As a result, users can focus on the logics of their applications rather on the intricacies of the underlying platforms.

    See project
  • infinimesh

    infinimesh is a opinionated multi-tenant hyperscale Internet of Things platform to connect IoT devices fast and securely with minimal TCO. It features a unique Graph-based authorization system, allowing users & engineers to create arbitrary hierarchical ontologies, with the possibility to scope permissions down to single sub-devices to specific users (e.g. suppliers). It exposes simple to consume RESTful & gRPC APIs with both high-level (e.g. device shadow) and low-level (sending messages)…

    infinimesh is a opinionated multi-tenant hyperscale Internet of Things platform to connect IoT devices fast and securely with minimal TCO. It features a unique Graph-based authorization system, allowing users & engineers to create arbitrary hierarchical ontologies, with the possibility to scope permissions down to single sub-devices to specific users (e.g. suppliers). It exposes simple to consume RESTful & gRPC APIs with both high-level (e.g. device shadow) and low-level (sending messages) concepts. The infinimesh IoT platform is open source and fully kubernetes compliant. No vendor lock-in - run it yourself on Kubernetes in your own datacenter, under your control with maximum data privacy.

    See project
  • Apache Hadoop

    -

    The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage. Rather than rely on hardware to deliver high-availability, the library itself is designed to detect and handle failures at the application layer, so delivering a highly-available service on top of a cluster…

    The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage. Rather than rely on hardware to deliver high-availability, the library itself is designed to detect and handle failures at the application layer, so delivering a highly-available service on top of a cluster of computers, each of which may be prone to failures.

    See project
  • Apache Flume

    -

    Flume is a distributed, reliable, and available service for efficiently collecting, aggregating, and moving large amounts of log data. It has a simple and flexible architecture based on streaming data flows. It is robust and fault tolerant with tunable reliability mechanisms and many failover and recovery mechanisms. It uses a simple extensible data model that allows for online analytic application.

    See project

View Alexander’s full profile

  • See who you know in common
  • Get introduced
  • Contact Alexander directly
Join to view full profile

Other similar profiles

Explore collaborative articles

We’re unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.

Explore More

Others named Alexander Alten