Introduction
At IBM, work is more than a job - it's a calling: To build. To design. To code. To consult. To think along with clients and sell. To make markets. To invent. To collaborate. Not just to do something better, but to attempt things you've never thought possible. Are you ready to lead in this new era of technology and solve some of the world's most challenging problems? If so, lets talk.
Your Role And Responsibilities
As Data Engineer, you will develop, maintain, evaluate and test big data solutions. You will be involved in the design of data solutions using Hadoop based technologies along with Python & Spark programming.
Responsibilities
- Responsible to Ingest data from files, streams and databases. Process the data with Spark, Scala, Kafka, Hive and scoop
- Develops Hadoop applications using Horton Works or other Hadoop distribution. Experienced with pulling data from various database systems, Network Elements and unstructured text from web, social media sites and other domain specific file
- Develop efficient software code for multiple use cases leveraging Python and Big Data technologies for various use cases built on the platform
- Provide high operational excellence guaranteeing high availability and platform stability
- Implement scalable solutions to meet the ever-increasing data volumes, using big data/cloud technologies Apache Spark, Kafka, any Cloud computing etc.
If you thrive in a dynamic, collaborative workplace, IBM provides an environment where you will be challenged and inspired every single day. And if you relish the freedom to bring creative, thoughtful solutions to the table, there's no limit to what you can accomplish here.
Required Technical And Professional Expertise
- Minimum 4+ years of experience in Big Data technologies
- Minimum 4+ years of experience in Python and Scala programming
- Experience in developing applications on Big Data and Cognitive technologies including API development
- Application Development background along with knowledge of Analytics libraries, open-source Natural Language Processing, statistical and big data computing libraries
- Expertise in Spark, Scala and Kafka technologies
- Ability to demonstrate micro / macro designing and familiar with Unix Commands and basic work experience in Unix Shell Scripting
- Demonstrated ability in solutioning covering data ingestion, data cleansing, ETL, data mart creation and exposing data for consumers
Preferred Technical And Professional Expertise
- Expertise in Python or Scala programming
- You love collaborative environments that use agile methodologies to encourage creative design thinking and find innovative ways to develop with cutting edge technologies
- Ambitious individual who can work under their own direction towards agreed targets/goals and with creative approach to work
- Intuitive individual with an ability to manage change and proven time management
- Proven interpersonal skills while contributing to team effort by accomplishing related results as needed
- Up-to-date technical knowledge by attending educational workshops, reviewing publications