This week - Remote Bigdata jobs
  • RetailNext

    remote, United States - Are you interested in distributed systems or big data?  RetailNext is looking for engineers to help build our cloud analytics system. We pull in and process data from thousands of brick and mortar stores to help our customers better understand and serve their…

  • Avocado Green Brands
    Preferred timezone: UTC -9 to UTC -1

    Avocado is an innovative lifestyle and eco-luxury brand, fueled by green living, sustainability, authenticity, social responsibility and remarkable customer experiences. We make mattresses and bedding products by hand in California using the finest natural and organic materials.

    We're seeking a talented and energetic Senior Software Developer - proficient in Elixir or FP languages - to join our tight-knit team. You must be focused on producing high quality and maintainable code. This role offers a high level of creativity, and an opportunity to design the future with a company that is challenging the status quo.

    Requirements:

    • 7+ years developing large-scale web / e-commerce applications

    • Production experience with Elixir/Phoenix, or a similar FP language

    • Full stack capabilities with a focus on performance, security and maintainability

    • Ability to collaborate and work well on a small agile team

    • Passionate, self motivated, pragmatic, goal-oriented, sense of humor :P

    • Infrastructure/server management (AWS) a plus

    • Experience with React a plus

    • Experience with data warehousing a plus

    • Experience with big data a plus

    Perks:

    • Fun, innovative culture

    • Ability to work remotely or on location at our offices in Hoboken and DTLA

    • Health, dental, vision

    • Travel and accommodations for technology / team building events

    • Company laptop

This month - Remote Bigdata jobs
  • BHE
    Must be located: United States of America.

    BHE’s mission is to improve healthcare through innovative analytics. Our team has built a next-generation analytics platform, Instant Health Data (IHD), to bring researchers together to generate insights into improving population health, quality of care, and managing costs.  

    Our engineering team is looking for talented individuals who want to bring our platform to the next level, expanding into new markets.  In undertaking this challenge, our new engineers will be at the forefront of the latest technologies, learn about large-scale, new and emerging data sources, and help BHE maintain its leadership position.

    Our engineers work in a fast-paced, rapid-learning environment with leaders in software development, working with massive data sets, and healthcare data analytics.  In our environment, everyone is encouraged to make a difference without experiencing the fixed ways of doing business in larger, bureaucratic organizations.   

    Job Description

    • Design, build, and maintain a highly scalable web analytics platform
    • Ensure that the platform meets business requirements and industry practices for security and privacy
    • Integrate new technologies and software engineering tools into existing platform
    • Mentor other software engineers
    • Provide software architecture support

    Minimum qualifications

    • Bachelor’s degree in Computer Science, Engineering, Math, or related technical/science field
    • 5 years of full-stack development experience
    • 5 years of experience working in a Linux environment

    Preferred qualifications

    • Significant experience with a NoSQL database
    • Significant experience with Apache Spark
    • Significant experience with Python and Java

    Why be a part of BHE's Team?

    • Leading healthcare data analytics/big data company
    • Work on a team of talented and pragmatic engineers/researchers
    • Great mentorship and growth opportunities
  • BookingSync

    BookingSync is more than a vacation rental solution for hosts and agencies; it is a platform for hundreds of applications and developers. We need an excellent engineer to take on the challenges of developing integration with multiple APIs from our partners, beautifully designed micro/macro-services, modeling complex domain and doing all that following TDD and writing a code of excellent quality.

    We are growing fast and so are our requirements for scalability and stability. We are no longer at a scale where adding X number of servers solves all the issues. Database sharding, replicating data via message queues/distributed streams, optimizing response times in APIs to the max, integration with multiple services coming from our own ecosystem and numerous third-party services, and very soon dealing with big data - those are the things you can surely expect to be challenged by.

    We are also a fully-remote async/distributed company across the entire world without a physical office. If traveling is your passion and you highly value location-independence, BookingSync is the right place for you.

    2200€ - 4500€ + VAT / month on B2B contract + equity (real shares, not stock options)

    You will need to have:

    • Strong experience with Ruby on Rails
    • Proficiency in backend development in general
    • Solid design and architecture skills
    • Ability to demonstrate that with the right approach Rails can scale easily
    • Ability to communicate effectively with both internal and external development teams
    • Experience with relational databases and SQL (we're on Postgres). Experience with other databases will be a plus
    • A strong commitment to quality
    • Excellent written and spoken English
    • Experience with remote work and/or a solid work ethic that makes it possible to thrive in a remote-only environment
    • Care for the greater good, BookingSync is not just a software company, we give back to humanitarian, solidarity, environmental & ecological causes from 10% to 50% of our profit.

    It would be great if you had experience with:

    • Messaging solutions (RabbitMQ, Kafka, etc.)
    • DDD/CQRS/ES, Clean Architecture
    • Contributing to development community through code, documentation, mentoring, teaching, speaking and all other forms
    • Short terms rentals or travel industry

    Sounds interesting? Contact us now!

    Want to increase your chances of standing out? Include your open source contributions, blog posts and other publications you wrote, links to conference talks. You could even tell us about one thing or two you created that you are particularly proud of or the most challenging features you've worked on.

    The hiring process consists of four steps:

    1. A small homework: Imagine that you don't want to or just don't have enough time to go through entire Twitter, LinkedIn or Youtube feeds manually every day and you just want to have a quick way of getting the content under a given hashtag. Implement a Rails application where you can provide a list of hashtags in the input and as the result, a feed aggregating the content from Twitter, LinkedIn and Youtube feeds under these hashtags will be displayed (including headlines, videos, descriptions and whatever else that could be useful). Make the design flexible enough that would make it easy to add more sources in the future. Treat it as a real-world application (not just a prototype or a proof of concept) and send us a link to the repo.

    2. Technical interview with remote pair-programming session with CTO: the interview part shouldn't take too much time - we are mostly interested in ability to solve the actual problems and being able to figure out the best possible solution for that problem, not in knowing things can be easily googled and learned through experience, that's why the interview part will be focused mostly on open-ended questions about some non-trivial matters. The pair-programming session will take a bit longer, and it will involve some nice challenge in a Rails universe.

    3. An interview with our CEO and/or COO that will be focused more on soft things, nothing technical this time.

    4. An opportunity to have the conversation with the entire team you are going to work with. It is an excellent way for both sides to make sure we will be a perfect mutual match.

  • Surge
    Must be located: North America.Preferred timezone: UTC +8

    Surge Forward is looking for a smart, self-motivated, experienced, senior-level remote developer to work as a long-term independent contractor.

    Experience Required: 

    Big Data, Hadoop EcoSystem, AWS

    Must be located in the US or Canada to be considered for this role. Sorry, No Visas.

    For immediate consideration, email resume with tech stack under each job and include what versions of Angular you have coded in (directly on the resume) as well as cell phone number and start date.

  • Metabolic Living
    PROBABLY NO LONGER AVAILABLE.

    Whats the Big Opportunity? This is your opportunity to join our team of 50+ to help us build a state-of-the-art B2C e-commerce platform & brand from the comfort of your own home for a rapidly scaling, profitable, 8-figure company. Were fully hitting our stride in the exploding $250+ billion health and wellness field. And this is your chance to help make a meaningful impact on the lives of people who struggle with their health by accelerating our ability to penetrate this market with our robust product offerings (see more below). To that end, our goal is to build the first billion-dollar brand thats 100% remote and need your help. To summarize the key points:

    • Salary: $110,000 - $140,000
    • Location: 100% Remote
    • Benefits: Full Insurance (health, dental & long-term disability) + Robust 401k
    • Hours: 8:30 a.m. 5:00 (and very flexible)
    • Culture: High-Performance but Humble (no jerks)

    Who are We Looking For? A full-time, senior, full-stack, brilliant, business-savvy web architect / web developer with heavy big data experience in the direct-to-consumer e-commerce space. Its a full-time, 100% remote position with a flexible work/vacation schedule as lifestyle is hugely important to our company. A curious mind, desire for autonomy while excelling in an open, truly collaborative culture of high performers makes would make you a perfect fit. These arent talking points. This is what we do, now. Were on pace to double this year and repeat that next year. To that end, we need more help tackling some very big engineering projects so we can handle our next leg of growth. Were laser-focused on recruiting someone with a business mindset who loves to solve big business problems by developing beautiful, efficient code and who enjoys having immediate, visible influence on the companys results. The ideal person is extremely results-driven and gets tremendous satisfaction from having their work see the light of day quickly and being motivated by seeing the impact it has. We want someone of who obsesses about and loves to code but also someone who loves to solve big business problems. Projects include everything from handling the massive system and data load that spending $1,000,000+ monthly on paid advertising brings, to building automated workflow tools to streamline internal operations, to developing sophisticated if/then logic in our billing code to handle the creative pricing and offers we use in our customer acquisition funnels. For example, we built our own all-in-one e-commerce system in 2012 as nothing handled the complex billing and funnel logic we needed to support our sophisticated B2C marketing strategies. However, now were moving over to the Shopify platform and strategically integrating our existing platform. That requires a complex analysis on which system does which part the best and building a smart integration. As another example, we build our own custom dashboards to precisely track our advertising efforts. To that end, we integrate with Google and Facebooks APIs to ingest all the data they provide. As a result, we end up with millions of rows of data a month that needs to be efficiently organized so we can use our BI tool to extract key business insights effectively. The ideal person is extremely self-motivated, loves to grind out work consistently and is comfortable being hands-on in all facets of web engineering from the browser-to-the-code-to-the-server. Experience with marketing funnels, e-commerce systems, split testing, sales analytics, conversion rate optimization is essential. We empower our developers with a lot of autonomy and responsibility for some very meaty projects that have very large $ impacts. To that end, this opportunity is suited only for a high-performing full-stack developer with an integrative mindset and proven ability to get things done cleanly and on time. You will work hand-in-hand with our CTO to optimize, support and manage our entire web infrastructure. You and our two other senior devs form our core development team of 4, covering for each other as vacation and other needs arise, while working synergistically on separate projects. Just as we obsess about creating metabolism-optimizing products that get the most results in the least time, were looking for someone equally obsessed about doing the same thing with efficient code. While your coding skills are important, it's how fast you process information, how quickly (and accurately) you can cycle through work and how well you work with the rest of our team that's crucial. To that end, communication is key to this role, as is the ability to self-motivate and quickly turn a business conversation into an actionable set of dev tasks that you can jump on immediately. Who is This Opportunity Perfect for? Anybody who truly is a top coder and creative architect who might relate to any of these: If youre tired of the commute, sick of burning the midnight oil, stress is dominating your life, and the cost of real estate is eating up your retirement, this is your opportunity to take back control, live life under your own terms and design a lifestyle that brings you joy, happiness and satisfaction. Or, if youve always wanted to be a part of the thrill of a rapidly growing company, but didnt want to move to an expensive big city with all the madness, stress and high cost that brings, this is your opportunity to stay where you are (or move to where you always wanted to go) to design your ideal life. Or, if youre drowning in the corporate world, suffocating from constantly shifting priorities that nobody agrees on, incessant scope creep, micromanaging bosses, endless meetings, constant interruptions and no respect for the quiet time you need to actually get work done, this is your opportunity to simplify and go deep to pour your heart and soul into ONE big opportunity and scale it like never before. Regardless of your situation, whether you prefer to work in the quiet of your own home, sipping a delicious cup of coffee at your favorite coffee shop, or outside enjoying the sunshine on a bright, beautiful day AND you are a humble, high-performing individual who would THRIVE in this type of environment, then please read on, as we cant wait to meet you.

    Benefits

    What You'll Enjoy from Us:

    • Competitive Compensation This respects the fact that you are a needle in a haystack, a true high-performer who any company would die for to have on their team.

    • Flexible Benefits While we offer an excellent benefits package, we can customize it to your specific situation, rather than forcing you to accept something that may not make sense for you.

    • 100% Remote A fantastic remote lifestyle with flexible hours when needed with ability to carve out time to do what you love.

    • Curated Team Peace of mind knowing we only recruit high-performing, humble people and that youll be surrounded by a supportive, high-energy team that respects you and your work.

      If this sounds like a good fit, if youre excited to do something challenging and meaningful that preserves your ability to enjoy life to the fullest, then please apply ASAP as we cant wait to meet you!

  • Ahrefs
    PROBABLY NO LONGER AVAILABLE.Preferred timezone: UTC -7 to UTC -3

    What We Need

    Ahrefs is looking for a Site Reliability Engineer to help take care of its distributed crawler powered by 2,000 servers and ensure all systems are up and running 24/7. If you possess a healthy desire to automate everything while being able to quickly resolve urgent issues manually, then we want you! We strive to keep humans away from doing repetitive jobs that can be done by computers and focus instead on foreseeing problems and defining programmatic means to handle them.

    Our system is big part custom OCaml code and also employs third-party technologies - Debian, ELK, Puppet, Clickhouse, and anything else that will solve the task at hand. In this role, be prepared to deal with 25 petabytes storage cluster, 2,000 baremetal servers, experimental large-scale deployments and all kinds of software bugs and hardware deviations on a daily basis.

    Basic Requirements:

    • Deep understanding of operating systems and networks fundamentals
    • Practical knowledge of Linux userspace and kernel internals

    The ideal candidate is expected to:

    • Understand the whole technology stack at all levels: from network and user-space code to OS internals and hardware
    • Independently deal with and investigate infrastructure issues on live production systems including dealing with hardware problems and interact with datacenters
    • Develop internal automation - monitoring, setup, statistics
    • Have the ability to foresee potential problems and prevent them from happening. Apply first-aid reaction to infrastructure failures when necessary
    • Help developers with deployment and integration
    • Participate in on-call rotation
    • Make well-reasoned technical choices and take responsibility for it
    • Approach problems with a practical mindset and suppress perfectionism when time is a priority
    • Setup automatic systems to control infrastructure
    • Possess a healthy detestation for complex shell scripts
  • PowerInbox
    PROBABLY NO LONGER AVAILABLE.$90,000.00 - $120,000.00.Preferred timezone: UTC -5

    PowerInbox is looking for a Senior Data Engineer

    *This job is fully remote (only in the USA, though) with the option to work from out NYC office*

    If you join us, what will you do?

    Build and maintain a real-time big data pipeline and reporting system for powerinbox. The data pipeline will feed our AI and analytics platform. The reporting system will automatically distribute reports to recipients on a configurable schedule. As needed, you will provide special reports as requested by sales and operations teams. This role offers opportunities to work with big data, data science, cloud computing, and the latest software technology.

    Specific Goals

    • Build and maintain a data pipeline for powerinbox machine learning.
    • Assist with the development of a data warehouse on which reports are derived.
    • Process 8 billion event transactions each month.
    • Assure data is captured and stored without loss.
    • Write code to provide reports for powerinbox.
    • Write a system that will run reports on a configurable schedule.
    • Increase revenue per 1,000 items by $0.10 each quarter.
    • Respond to ad-hoc requests for information.

    In order to be great at your job,

    You Are

    A fast learner; have great analytical skills; relentless and persistence in accomplishing goals; enthusiastic with an infectious personality.

    You Work

    Efficiently; with flexibility; proactively; with attention to detail; to high standards.

    Together We

    Emphasize honesty and integrity; require teamwork; have open communication; follow-through on commitments; stay calm under pressure.

    You Have

    Four to six years experience with Python or R; three or more years experience developing and deploying software on Linux; three or more years working with SQL; at least two years experience providing data analysis; professional experience with data science knowledge; and working knowledge of BI tools and software.

    This is extra, but if you have it, it will make us happy

    • Experience working remotely
    • Knowledge of/interest in the digital and AdTech landscape
    • Experience working with big data

    About PowerInbox

    Who We Are

    We are a digital monetization startup ecosystem that is always open to new talent

     Why We Are

    Personalization is key and we at PowerInbox believe that email is not meant to be stationary and static, but relevant and filled with dynamic content and advertisements.

    What We Are

    We at PowerInbox boost your revenue and brand engagement through real-time advertising, and native ad displays.

     If interested please send your resume to [email protected]

  • Nagarro
    PROBABLY NO LONGER AVAILABLE.Preferred timezone:

    Required experience and skills: 

    • Expertise in Java or Scala

    • Familiarity with cluster computing technologies such as Apache Spark or Hadoop MapReduce

    • Familiarity with relational and big data such as Postgres, HDFS, Apache Kudu and similar technologies

    • Strong skills in analytic computing and algorithms

    • Strong mathematical background, including statistics and numerical analysis

    • Knowledge of advanced programming concepts such as memory management, files & handles, multi-threading and operating systems.

    • Passion for finding and solving problems

    • Excellent communication skills, proven ability to convey complex ideas to others in a concise and clear manner 

    Desirable experience and skills: 

    • Familiarity with scripting languages such as Python or R

    • Experience in performance measurement, bottleneck analysis, and resource usage monitoring

    • Familiarity with probabilistic and stochastic computational techniques

    • Experience with data access and computing in highly distributed cloud systems

    • Prior history with agile development

Older - Remote Bigdata jobs
  • Dyanata
    PROBABLY NO LONGER AVAILABLE.Preferred timezone: UTC -11 to UTC -3

    Provides expertise in the definition, adoption and adherence to enterprise business intelligence and information architecture strategies, processes and standards. Responsible for delivering enterprise information architecture and technical / infrastructure architecture solutions for the enterprise data warehouse and end user reporting tools. This position provides technology direction, assistance and training to teams performing BI-related activities, both inside and outside IT Services. Will Support the development of the global BI strategic road map by conducting market analysis and technology reviews, and establishing strong partnerships with key vendors to enable road map implementation.

    If you feel that AWS S3 is a great solution as a datalake and leveraging EMR/Spark for data transformations into a Redshift data warehouse.  Then Critical Mix is a great fit.

    Office Locations

     Perrysburg, Oh -  Dallas, Tx - Warwick, RI - San Francisco, CA

    Skills & Requirements

    Responsible for designing and maintaining complete data systems architecture for the Critical Mix organization on enterprise levels

    • Develop and Implement road maps and phased plans that gradually increases the enterprise data architecture maturity level of Critical Mix over a multi-year period and support both current & future state business processes
    • Sets the overall direction, policies and guidelines for the Enterprise Data Architecture function
    • Maintains an ongoing partnership with the business to apply in-depth knowledge of the business operations, strategies, priorities and information requirements to establish the technical direction at an enterprise view
    • Promotion and delivery of shared infrastructure and applications to reduce costs and improve information flows
    • Defines the system, technical, and application architectures, and in some instances the business systems/process architecture for major areas of development
    • Creates processes and standards to develop, maintain and integrate enterprise architectures within the process of strategy development and technology planning
    • Ensures appropriate technical standards, procedures, and governance are defined and followed
    • Ensures enterprise solutions are scalable and adoptable according to changing business needs
    • Maintains close working relationships with executive and business management to understand the Critical Mix strategy and requirements
    • Keep up to date with new technologies and actively implement innovative solutions

    Requirements

    • Bachelor's degree required; preferably in MIS
    • 8+ years experience in IT including experience with systems architecture methodologies
    • 3+ years of hands-on development experience in more than one of these technologies: Oracle, SQL Server, NoSQL, Big Data Technologies like Hadoop, Redshift, Snowflake or other DW storage architecture, Chartio or similar, Informatica or similar.
    • Dimensional data modeling principles (star and snowflake schemas, denormalized data structures, slowly changing dimensions, etc.)
    • Physical data architectures: data warehouses, independent data marts
    • Data integration tools: ETL (extract, transform, load), CDC (change data capture)
    • Data quality, master data management, metadata management, collaboration and business process management Bachelors or above degree in Computer Science, or related discipline, or equivalent experience.
  • Supermercato24
    PROBABLY NO LONGER AVAILABLE.

    Who are we looking for:

    We are looking for a savvy Data Engineer to join our growing tech team.

    You will support our software developers, data analysts and data scientists on data initiatives and will ensure optimal data delivery architecture is consistent throughout ongoing projects.

    The right candidate will be excited by the prospect of optimizing or even re-designing our company’s data architecture to support our next generation of products and data initiatives.

    Roles and Responsibilities:

    • you will build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and other big data technologies
    • you will design and model data structures to help analyzing our business and technical data
    • you will support existing processes running in production
    • you will work together with people from other key areas to assist with data-related technical issues and support their data infrastructure needs

    Skills & Requirements

    • knowledge in relevant engineering best practices, data management fundamentals, data storage principles, and be current with recent advances in distributed systems as it pertains to data storage and computing
    • 2+ years of experience in designing, building and maintaining data architecture(s) and infrastructure(s), both relational and non-relational
    • 2+ years of maintaining data warehouse systems and working on large scale data transformation using SQL, Hadoop, Hive, or other Big Data technologies; experience with ETL tools is a plus
    • 2+ years of data modeling experience, and able to use data models to improve the performance of software services
    • experience with Cloud Based Solution (AWS Redshift, GCP Big Query) and programming language (Python, Java) is a plus
    • experience communicating with colleagues from engineering, analytics, and business backgrounds
    • degree in Engineering, Math, Statistics, Computer Science, or related discipline or equivalent experience is a plus.
    • be able to legally work in Europe (you are the holder of a EU Passport or you are the holder of EU residency permit or you are the holder of a Schengen Work Visa)
  • APEX Expert Solutions
    PROBABLY NO LONGER AVAILABLE.$110,000.00 - $140,000.00.Preferred timezone: UTC -9 to UTC -1

    Duties and Responsibilities

    • Work in an fast-paced agile development environment architecting and developing Hadoop applications
    • Provide technology recommendations for potential product application development
    • Gather and analyze requirements from product owners ensuring products meet business requirements
    • Collaborate with other software engineers and team leads in designing and developing software solutions which meet high quality standards
    • Quickly prototype and develop Python/Java/Scala applications in diverse operating environments capable of interfacing with NoSQL datastores such as Accumulo and HBase
    • Write efficient code to extract, transform, load, and query very large datasets to include both structured and unstructured datasets
    • Develop standards and new design patterns for Big Data applications and master the tools and technology components within the Hadoop and Cloudera environments
    • Design and implement REST API applications provide web application connectivity to backend datastores

    Skills & Requirements

    • 3 years of building Java applications including framework experience (J2EE, Spring, etc.)
    • 1 year of building and coding applications using Hadoop components – HDFS, HBase, Hive, Sqoop, Flume, Spark, etc
    • 3 years experience with Spark
    • 1 year of experience with GeoMesa
    • 1 year of experience with SparkSQL
    • Experience building and maintaining Cloudera-based clusters
    • Experience using traditional ETL tools & RDBMS
    • Experience developing REST web services
    • Demonstrated effective and successful verbal and written communication skills
    • Bachelor degree in Computer Science or related technological degree
    • U.S. citizen

    Desired Qualifications

    • Full life cycle software application development experience
    • Front end web development with experience in JQuery, Polymer, web components, Bootstrap, Node.js, etc
    • Demonstrated ability to quickly learn and apply new technologies
    • Experience with unstructured datasets, such as log files, email, text
    • Experience with geospatial datasets and datastores
  • Surge
    PROBABLY NO LONGER AVAILABLE.Must be located: North America.Preferred timezone: UTC -8

    SURGE is looking for smart, self-motivated, experienced, senior engineers who enjoy the freedom of telecommuting and flexible schedules, on a variety of software development projects.

    REQUIRED:

    Data Engineer Openings requiring ETL and Hadoop 

    Must be located in the US or Canada to be considered for this role. Sorry, No Visas.

    For immediate consideration, email resume with tech stack under each job and include your cell phone number and start date: [email protected]

  • Nagarro
    PROBABLY NO LONGER AVAILABLE.

    Required Experience and Skills

    • Expertise programming experience with Java
    • Expertise in Linux software development
    • Building observability into services via instrumentation, logging, and tracing.
    • Good understanding of microservices concepts/architecture
    • Experience developing cloud services and cloud platforms
    • Experience with agile development and working with agile engineering teams
    • Excellent communication skills, proven ability to convey complex ideas to others in a concise and clear manner
    • BS/MS in Computer Science, Computer Engineering, Math, or equivalent professional experience

    Desirables / Good to have:

    • Experience with Scala and Python
    • Experience with HDFS, Spark, and relational databases such as Postgres
    • Experience with open source monitoring and logging technologies such as Prometheus and ELK
    • Experience implementing containers in a microservices environment
    • Experience educating engineers about what to log, measure and alert on with an emphasis on surfacing trends to be used for SLA/O analysis
    • Experience in building analysis tooling and processes to provide mature mechanics to help users understand the state of a distributed system at a given point in time.
    • Experience in creating and delivering performance monitoring and insights for bespoke platforms via dashboards, scorecards and ad hoc analysis.
  • Critical Mix
    PROBABLY NO LONGER AVAILABLE.

    Provides expertise in the definition, adoption and adherence to enterprise business intelligence and information architecture strategies, processes and standards. Responsible for delivering enterprise information architecture and technical / infrastructure architecture solutions for the enterprise data warehouse and end user reporting tools. This position provides technology direction, assistance and training to teams performing BI-related activities, both inside and outside IT Services. Will Support the development of the global BI strategic road map by conducting market analysis and technology reviews, and establishing strong partnerships with key vendors to enable road map implementation.

    If you feel that AWS S3 is a great solution as a datalake and leveraging EMR/Spark for data transformations into a Redshift data warehouse.  Then Critical Mix is a great fit.

    Office Locations

     Perrysburg, Oh -  Dallas, Tx - Warwick, RI - San Francisco, CA

    Skills & Requirements

    Responsible for designing and maintaining complete data systems architecture for the Critical Mix organization on enterprise levels

    • Develop and Implement road maps and phased plans that gradually increases the enterprise data architecture maturity level of Critical Mix over a multi-year period and support both current & future state business processes
    • Sets the overall direction, policies and guidelines for the Enterprise Data Architecture function
    • Maintains an ongoing partnership with the business to apply in-depth knowledge of the business operations, strategies, priorities and information requirements to establish the technical direction at an enterprise view
    • Promotion and delivery of shared infrastructure and applications to reduce costs and improve information flows
    • Defines the system, technical, and application architectures, and in some instances the business systems/process architecture for major areas of development
    • Creates processes and standards to develop, maintain and integrate enterprise architectures within the process of strategy development and technology planning
    • Ensures appropriate technical standards, procedures, and governance are defined and followed
    • Ensures enterprise solutions are scalable and adoptable according to changing business needs
    • Maintains close working relationships with executive and business management to understand the Critical Mix strategy and requirements
    • Keep up to date with new technologies and actively implement innovative solutions

    Requirements

    • Bachelor's degree required; preferably in MIS
    • 8+ years experience in IT including experience with systems architecture methodologies
    • 3+ years of hands-on development experience in more than one of these technologies: Oracle, SQL Server, NoSQL, Big Data Technologies like Hadoop, Redshift, Snowflake or other DW storage architecture, Chartio or similar, Informatica or similar.
    • Dimensional data modeling principles (star and snowflake schemas, denormalized data structures, slowly changing dimensions, etc.)
    • Physical data architectures: data warehouses, independent data marts
    • Data integration tools: ETL (extract, transform, load), CDC (change data capture)
    • Data quality, master data management, metadata management, collaboration and business process management Bachelors or above degree in Computer Science, or related discipline, or equivalent experience.
  • phData
    PROBABLY NO LONGER AVAILABLE.

    If you're inspired by innovation, hard work and a passion for data, this may be the ideal opportunity to leverage your background in Big Data and Software Engineering, Data Engineering or Data Analytics experience to design, develop and innovate big data solutions for a diverse set of global and enterprise clients.  

    At phData, our proven success has skyrocketed the demand for our services, resulting in quality growth at our company headquarters conveniently located in Downtown Minneapolis and expanding throughout the US. Notably we've also been voted Best Company to Work For in Minneapolis for the last 2 years.   

    As the world’s largest pure-play Big Data services firm, our team includes Apache committers, Spark experts and the most knowledgeable Scala development team in the industry. phData has earned the trust of customers by demonstrating our mastery of Hadoop services and our commitment to excellence.

    In addition to a phenomenal growth and learning opportunity, we offer competitive compensation and excellent perks including base salary, annual bonus, extensive training, paid Cloudera certifications - in addition to generous PTO and employee equity. 

    As a Solution Architect on our Big Data Consulting Team, your responsibilities will include:

    • Design, develop, and innovative Hadoop solutions; partner with our internal Infrastructure Architects and Data Engineers to build creative solutions to tough big data problems.  

    • Determine the technical project road map, select the best tools, assign tasks and priorities, and assume general project management oversight for performance, data integration, ecosystem integration, and security of big data solutions.  Mentor and coach Developers and Data Engineers. Provide guidance with project creation, application structure, automation, code style, testing, and code reviews

    • Work across a broad range of technologies – from infrastructure to applications – to ensure the ideal Hadoop solution is implemented and optimized

    • Integrate data from a variety of data sources (data warehouse, data marts) utilizing on-prem or cloud-based data structures (AWS); determine new and existing data sources

    • Design and implement streaming, data lake, and analytics big data solutions

    • Create and direct testing strategies including unit, integration, and full end-to-end tests of data pipelines

    • Select the right storage solution for a project - comparing Kudu, HBase, HDFS, and relational databases based on their strengths

    • Utilize ETL processes to build data repositories; integrate data into Hadoop data lake using Sqoop (batch ingest), Kafka (streaming), Spark, Hive or Impala (transformation)

    • Partner with our Managed Services team to design and install on prem or cloud based infrastructure including networking, virtual machines, containers, and software

    • Determine and select best tools to ensure optimized data performance; perform Data Analysis utilizing Spark, Hive, and Impala

    • Local Candidates work between client site and office (Minneapolis).  Remote US must be willing to travel 20% for training and project kick-off.

    Technical Leadership Qualifications

    • 5+ years previous experience as a Software Engineer, Data Engineer or Data Analytics

    • Expertise in core Hadoop technologies including HDFS, Hive and YARN.  

    • Deep experience in one or more ecosystem products/languages such as HBase, Spark, Impala, Solr, Kudu, etc

    • Expert programming experience in Java, Scala, or other statically typed programming language

    • Ability to learn new technologies in a quickly changing field

    • Strong working knowledge of SQL and the ability to write, debug, and optimize distributed SQL queries

    • Excellent communication skills including proven experience working with key stakeholders and customers

    Leadership

    • Ability to translate “big picture” business requirements and use cases into a Hadoop solution, including ingestion of many data sources, ETL processing, data access and consumption, as well as custom analytics

    • Experience scoping activities on large scale, complex technology infrastructure projects

    • Customer relationship management including project escalations, and participating in executive steering meetings

    • Coaching and mentoring data or software engineers