Data Engineer

Bill.Com Palo Alto , CA 94306

Posted 3 months ago

About is a leader in financial process automation for small businesses and mid-size companies. Making it simple to connect and do business, the Back Office Cloud digitizes, automates and simplifies legacy payment and financial processes. With an integrated, end-to-end platform, leverages artificial intelligence to reduce manual work, and provides a cloud workspace to help run your business anytime, anywhere. The company partners with many of the largest U.S. financial institutions, more than 70% of the top 100 U.S. accounting firms, and major accounting software providers. manages more than $70B in annual payment volume across ACH, virtual cards, checks, and international payments. The company has offices in Palo Alto, California and Houston, Texas. For more information, or follow @billcom.

Mission: moves over $60B per year and we have 10 years worth of customer data. We are leveraging this data to make data driven decisions, and apply data science and machine learning to solve a variety of tough problems. We are in the middle of a large-scale transformation to the public cloud and are developing data pipelines, data warehouse, and machine learning infrastructure in AWS.

Data engineers at will be responsible for building data pipelines and the infrastructure to enable data science, data analytics, and machine learning at scale in AWS. Some of the problems we are currently working on include: detecting payment fraud, extracting semantic data from customer documents, and increasing customer acquisition through advanced analytics. Data engineers will own and build the data platform that makes all of this possible. We have multiple positions available at different levels of seniority.

Professional Experience/Background to be successful in this role:

  • 5+ years of experience owning and building data pipelines.

  • Extensive knowledge of data engineering tools, technologies and approaches

  • Ability to absorb business problems and understand how to service required data needs

  • Design and operation of robust distributed systems

  • Proven experience building data platforms from scratch for data consumption across a wide variety of use cases (e.g data science, ML, scalability etc)

  • Demonstrated ability to build complex, scalable systems with high quality

  • Experience with multiple data technologies and concepts such as Airflow, Kafka, Hadoop, Hive, Spark, MapReduce, SQL, NoSQL, and Columnar databases.

  • Experience with specific AWS technologies (such as S3, Redshift, EMR, and Kinesis) a plus

  • Experience in SQL and one or more of Python, Java and Scala

Expected Outcomes:

  • Design and implement data infrastructure and processing workflows required to support data science, machine learning, BI and reporting in AWS

  • Build robust, efficient and reliable data pipelines consisting of diverse data sources

  • Design and develop real time streaming and batch processing pipeline solutions

  • Own the data expertise and data quality for the pipelines

  • Drive the collection of new data and refinement of existing data sources

  • Identify shared data needs across, understand their specific requirements, and build efficient and scalable pipelines to meet various needs

  • Build data stores for feature variables required for machine learning Culture:

Humble No ego

Fun Celebrate the moments

Authentic We are who we are

Passionate Love what you do

Dedicated To each other and the customer

icon no score

See how you match
to the job

Find your dream job anywhere
with the LiveCareer app.
Mobile App Icon
Download the
LiveCareer app and find
your dream job anywhere
App Store Icon Google Play Icon

Boost your job search productivity with our
free Chrome Extension!

lc_apply_tool GET EXTENSION

Similar Jobs

Want to see jobs matched to your resume? Upload One Now! Remove
Data Engineer


Posted 2 months ago

VIEW JOBS 8/21/2019 12:00:00 AM 2019-11-19T00:00 About Us: We are The Selling App. The fast and easy way to sell or buy almost anything, from fashion to toys, sporting goods to electronics, jewelry to shoes. Launched in 2013, Mercari quickly became the #1 shopping app in Japan. Now we're on a global mission to build a future where people everywhere feel empowered to sell the things they don't use, a future where all useful things are used. And with a fast-growing user base in the U.S. of over 45 million downloads, we are on our way to doing just that. The ideal candidate is eager to take responsibilities on data products and machine learning platform spontaneously and has a passion to make our systems reliable and sustainable with the modern technologies. The candidate will take responsibility for not only data engineering, but also machine learning platform. We expect the candidate is able to design and implement data pipelines at scale and is deeply familiar with container technologies and related topics to offer intelligent system as microservices. What You'll Be Doing: * Design, build and operate ETL pipeline at scale. * Automation of processes related to data products and machine learning products. * Support stakeholders with designing data structure for data products. * Development and operation of API/tools related to data products and machine learning products. What You'll Need: * Industry experience building and productionizing data pipelines. * Solid understanding / experience in the machine learning space. * Industry experience building and productionizing system to serve machine learning. * Ability and desire to take full ownership of projects, driving them forward to completion. * Experience with Google Cloud Platform or related cloud services. Nice-to-haves: * Experience with Apache Spark, Apache Beam or related distributed processing frameworks. * Experience with Apache Airflow or related workflow scheduling products. * Experience with developing microservices with docker and kubernetes to serve machine learning models. Technologies We Use: * ETL: Apache Airflow * Container: Docker/Kubernetes * API: gRPC/Tensorflow Serving/Flask(REST) * Database: Google Datastore/MySQL/Google Spanner * Distributed Processing: Apache Beam/Apache Spark, Hadoop * Cloud: Google Cloud(BigQuery/ML Engine/Google Dataflow/Google Dataproc, etc.) Why Mercari? Mercari nurtures an all for one environment where teamwork and innovative thinking is the priority. Perks: * Competitive medical, dental, and vision insurance options * 401k match * Life & disability insurance * Employee Assistant Program * New parent paid leave * Rocket Lawyer legal services * Fond perks and rewards * Commuter reimbursement * Time when you need it - unlimited vacation days * Catered lunches everyday * Team outings and events Mercari Palo Alto CA

Data Engineer