这是嘛意思啊。# Piebridge - 鹊桥
g*r
1 楼
Company: Okta
Location: SF or San Jose
如果感兴趣,请发简历到 [email protected]
https://www.okta.com/company/careers/ga/senior-data-engineer--983309/
Senior Data Engineer
Position Description:
We are looking for an exceptional Senior Data Engineer who is responsible
for designing and developing robust, scalable solutions for large-scale data
infrastructure in a fast-paced agile environment. You will be encouraged to
initiate new projects to collect, store, and analyze data and to partner
with analysts, engineers, business stakeholders to push Okta forward. The
ideal candidate will have a strong engineering background with the ability
to tie engineering initiatives and business impact. You will participate in
detailed technical design, development and implementation of applications
using cutting-edge technology stacks.
Job Duties and Responsibilities:
Design highly-scalable and reliable data pipelines to consume, integrate and
analyze large volumes of complex data from different sources (both batch
and near-real-time) to support the fast-growing needs of our business
Build data access layer to provide data service to internal and external
business stake holders
Constantly evolve data pipeline and data models to balance scalability and
performance
Strong understanding of analytics needs and proactively come up with
architectural patterns to improve the efficiency
Interface with Data Analysts and Data Scientists to gather requirements and
build functionality
Support and enhance existing data infrastructure
Build data expertise and own data quality for allocated areas of ownership
Build and experiment with different tools and technology, and share
learnings with the team
Contribute to evaluation of new technologies
Required Skills:
BS Computer Science, Engineering or another quantitative field of study
5-7 years in a data engineering role
Strong knowledge in data warehouse and data lake space
Experience with building distributed system
Expertise in a programming language (Preferably Python or JAVA)
SQL, SQL, SQL!
Strong Experience with ETL tools such as Pentaho, Talend, or Informatica
Strong experience with relational database and columnar MPP database
Redshift or Teradata, and Vertica
Strong experience with database and application performance tuning
Experience with different operating system (Unix/Linux/Windows)
Experience with Continuous Integration and Continuous Deployment tools and
procedures such as Jenkins, Git, Chef, and Ansible
Experience with cloud infrastructure/platform(AWS, Azure, Google Cloud
Platform)
Hands on and deep experience with schema design and data modeling
Familiar with Jira, Confluence and Agile methodology
Experience with real-time or near real-time data streaming using Storm or
Spark
Experience with Kinesis or Kafka, Storm, Elastic, S3, Splunk, Spectrum,
Redis, MongoDB, Spark, Zookeeper
Experience with building API using Java, python, php or other programming
languages
Familiar with Tableau, Looker, Qlikview, Microstrategy, or other data
visualization tools
Experience with CRM systems such as Salesforce
Preferred Skills:
Distributed data visualization web application development experience
Machine learning
Big Data concept and experience
Soft Skills:
Team player
Detail oriented
Self-driven personality
Innovative and ability to execute
Excellent oral and written communication skills. Must be able to interact
cross-functionally with both technical and non-technical people
Location: SF or San Jose
如果感兴趣,请发简历到 [email protected]
https://www.okta.com/company/careers/ga/senior-data-engineer--983309/
Senior Data Engineer
Position Description:
We are looking for an exceptional Senior Data Engineer who is responsible
for designing and developing robust, scalable solutions for large-scale data
infrastructure in a fast-paced agile environment. You will be encouraged to
initiate new projects to collect, store, and analyze data and to partner
with analysts, engineers, business stakeholders to push Okta forward. The
ideal candidate will have a strong engineering background with the ability
to tie engineering initiatives and business impact. You will participate in
detailed technical design, development and implementation of applications
using cutting-edge technology stacks.
Job Duties and Responsibilities:
Design highly-scalable and reliable data pipelines to consume, integrate and
analyze large volumes of complex data from different sources (both batch
and near-real-time) to support the fast-growing needs of our business
Build data access layer to provide data service to internal and external
business stake holders
Constantly evolve data pipeline and data models to balance scalability and
performance
Strong understanding of analytics needs and proactively come up with
architectural patterns to improve the efficiency
Interface with Data Analysts and Data Scientists to gather requirements and
build functionality
Support and enhance existing data infrastructure
Build data expertise and own data quality for allocated areas of ownership
Build and experiment with different tools and technology, and share
learnings with the team
Contribute to evaluation of new technologies
Required Skills:
BS Computer Science, Engineering or another quantitative field of study
5-7 years in a data engineering role
Strong knowledge in data warehouse and data lake space
Experience with building distributed system
Expertise in a programming language (Preferably Python or JAVA)
SQL, SQL, SQL!
Strong Experience with ETL tools such as Pentaho, Talend, or Informatica
Strong experience with relational database and columnar MPP database
Redshift or Teradata, and Vertica
Strong experience with database and application performance tuning
Experience with different operating system (Unix/Linux/Windows)
Experience with Continuous Integration and Continuous Deployment tools and
procedures such as Jenkins, Git, Chef, and Ansible
Experience with cloud infrastructure/platform(AWS, Azure, Google Cloud
Platform)
Hands on and deep experience with schema design and data modeling
Familiar with Jira, Confluence and Agile methodology
Experience with real-time or near real-time data streaming using Storm or
Spark
Experience with Kinesis or Kafka, Storm, Elastic, S3, Splunk, Spectrum,
Redis, MongoDB, Spark, Zookeeper
Experience with building API using Java, python, php or other programming
languages
Familiar with Tableau, Looker, Qlikview, Microstrategy, or other data
visualization tools
Experience with CRM systems such as Salesforce
Preferred Skills:
Distributed data visualization web application development experience
Machine learning
Big Data concept and experience
Soft Skills:
Team player
Detail oriented
Self-driven personality
Innovative and ability to execute
Excellent oral and written communication skills. Must be able to interact
cross-functionally with both technical and non-technical people