We convert our passion and experience into data-driven solutions. Our mission is to help organisations to address their data-processing challenges using popular open-source tools from the Hadoop Ecosystem and beyond.
We implement data-driven applications in frameworks like Spark, Hive, Storm, Flink and more.
We are experts at installation, configuration and operations of Hadoop clusters and various projects that integrate with it e.g. Spark, Kafka, Hive, HBase, Druid, Falcon.
Our high-quality training for Big Data administrators, engineers and analysts are delivered by our consultants who work with this technology on daily basis.
As consultants, we have been involved in multiple successful Big Data projects. Some of them includes:
We have installed a production Hadoop cluster, migrated an existing MySQL-based data infrastructure to the Hadoop ecosystem, implemented a wide range of ETL processes
We have builded and are maintaining a scalable real-time stream processing system to calculate KPIs and feed dashboarding system. Used technologies: Hadoop, Kafka, Spark
We delivered two training for Data Analysts (4 days) and Hadoop Administrators (3 days) for 15 people in total.
We have been also supporting numerous of undisclosed customers ranging from startup to global corporations to address their Big Data challenges. This includes consulting
We had been working with Big Data technologies as full-time employees at multiple well-known companies e.g. Spotify, Allegro, Netezza, Gadu-Gadu, the University of Warsaw. At GetIndata we have also helped numerous data-driven companies to benefit from their data even more. Some of them include:
We are data developers and system administrators with practical multi-year experience in Big Data technologies. Full engagement, true passion, continuous improvement and strong desire to challenge the status quo is a big part our DNA.
Piotr is a system administrator with passion for Open Source. He has gained his first experience with distributed systems working for companies like GarageFarm.NET and CopernicusComputing, where he helped to build and maintain HA infrastructure. Now he works through Hadoop ecosystem and supports colleagues from GetInData.
Adam became a fan of Hadoop after implementing his first MapReduce job in 2010. Since then he has been working with Hadoop at Netezza, the University of Warsaw, Spotify (where he had operated one of the largest and fastest-growing Hadoop clusters in Europe for two years), as an Authorized Cloudera Training Partner and finally now at GetInData.
Piotr has extensive practical experience in writing applications running on Hadoop, maintaining Hadoop clusters and delivering Big Data trainings for administrators, developers and analysts. At Spotify, he was part of the team operating arguably the biggest Hadoop cluster in Europe.
Adam joined our team as a Linux specialist. He has gathered experience working at Gadu-Gadu and financial companies, including banks. Managing thousands servers is nothing surprising for him. Building highly available and scalable systems to store and analyse data is his main goal.
Krzysztof is an architect, engineer and researcher of solutions that use Big Data technologies, like advanced analytics, decision automation systems or recommendation engines. He likes to work full-stack going from architecting solution through engineering down to installing, troubleshooting & monitoring. Now specializes in scalable real-time analytics solutions.
Klaudia is a student of MSc degree at Warsaw University of Technology. During gaining her Bachelor degree she acquired fundamental knowledge within creating DataBase. In GetInDate Klaudia uses her financial, human resources and public relations skills to support company in all managment operations.
We are looking for an engineer with true passion and practical experience in Hadoop (and/or other projects) who can join our expert Big Data team in Warsaw. If you are interested, read more.
We are looking for Big Data Administrator who loves Linux and operating large-scale distributed systems to join our expert Big Data team in Warsaw. If you are interested, read more.
We share our knowledge happily. Apart from writing blog posts, we also give technical reviews of books on distributed systems (e.g. about Hadoop, HBase, Impala), we speak international Hadoop conferences (e.g. Strata Hadoop World, Hadoop Summit Europe and Hadoop Summit San Jose) and talk at Hadoop meetups (e.g. San Diego, Stockholm, Warsaw, Krakow and more). Our members co-founded and co-organize Hadoop User Groups in Warsaw and Stockholm.
Camus, a MapReduce job that loads data from Kafka into HDFS, has a number of time-related